Sei sulla pagina 1di 46

Cognos Interview Questions

1.Explain how to create power play reports


To create a power play report
1)You have to create a IQD file from impromptu by saving IMR report as IQD
2)Use Frame Work manager model and externalize the query subject you want to use in powerplay
transformer and power play
When you have the IQD file published to a location such as UNC
Go to power play transformer model and select new, select the data source name, and IQD file location
either published from FM or Impromptu (saved as IQD) You will see all the query items in the explorer then
you can specify the dimensions like time and measures. Then generate categories for the dimensions and
measures .Right click the cube and select create power cube, after that view the cube in cognos power play
explorer. We can also publish the cube to PPES(Power play enterprise server)
Publish the cube to the upfront Use power play web explorer to view the cube
2. what type problems we can face in general at report running time
The most common problems are 1. No Data Appears in the Report (to remove this check data source or package)
2. Unexpected or incorrect values appears in reports, may be report is running with limited data
3. The report filter does not work; values are not coming in filter option
4. Report is not able to open in excel, cvs or xml
3.How can i test reports in cognos?
In cognos report net by the validate report option a report can be tested. If there will be any error, it will
specify the error, unless it will give message -'report specification is valid'.
4.How can i schedule reports in cognos
By using Cognos Scheduler, one can schedule the running reports in Impromptu to execute and save it in
desired format. By using Cognos MACRO script language the reports can be executed and distributed to
recipients by using mail applications. Compiled Cognos Macros can be scheduled using Cognos Scheduler.
5.What is the difference between a cascading report and drill thru report? Why do we go for drill thru
report?
Cascading report works based on the condition,but drill thru work based on the data item what we select as
a drill thru options.
6.What is meant by Junk Dimension?
A dimension which does not changes the grain level is called junk diamension.grain lowest level of reporting.
7. What is meant by Junk Dimension?How do you perfom while running the report?Where will you
see the time
Is LOOP CONSTRAINTS occur in COGNOS, if yes how to resolve that and please tell me how the loops
occur in cognos?
Problem Description
what are looping joins and how do you resolve them?
Solution Description
Looping joins could potentially return incorrect data.
An example of a looping join,
B -> C
A -> < > -> D
E -> F
When you select an item from table A and D, Impromptu will try to choose the shortest path, e.g. if A -> D
existed, then this is the path Impromptu will take. But in the above situation, the two paths are equal and so
Impromptu has to make a choice, e.g. "A -> B -> C -> D" OR "A -> E -> F -> D". Impromptu makes it's choice
based on how the catalog was constructed, which cannot be altered once it's created; order of the tables in
the catalog.
The two paths could return different results depending on the relationship between the tables in the path.

The report would be fine IF Impromptu chose the expected path. The choice is not always right.
Eliminating looping joins prevents Impromptu from making the wrong choice. To eliminate looping joins, you
can either break unnecessary joins, e.g. reports do not need a join between table F and D, e.g.
B -> C -> D
A -> <
E -> F
But if you need all the joins, use ALIAS tables to break the looping join. Add an alias table for table A and
break the join from table A and E, e.g.
A -> B -> C
> -> D
Alias A -> E -> F
Both solutions could affect existing reports.
Title:
Created:
Applies To:

Looped joins
Nov 04, 1999
Impromptu - 2.0
Impromptu - 3.01
Impromptu - 3.03
Impromptu - 3.04
Impromptu - 3.5
Impromptu - 4.0
Impromptu - 5.0
Impromptu - 7.1

Problem Description
Under the Joins dialog on the Analyze Tab it states a Loop join is present. What does this mean and how can
it be resolved?
Solution Description
A Loop Join occurs when there are multiple paths between database tables. An example of this is A joins to B
and B joins to C and C joins to A.
The proper definition of join strategies in an Impromptu catalog is crucial to the success of an ad-hoc reporting
environment. Impromptu shelters the user from having to know any of the technical information about the
database, including name, location, table and column names, and join strategies. The Impromptu
Administrator must be very thorough in their definition and testing of the join strategies. Impromptu provides
an ability to analyze the joins and determine any anomalies. The most common is the Loop Join.
The implications of the loop join are that there is no way to predetermine which of the various join paths will be
used by Impromptu when creating the SQL. SQL is dynamically generated for each report as it is created and
before it executes. For example, to create a report using columns from tables A and C, we could join from
A=>B=>C or directly from A=>C. In some cases, both of these joins would result in the same data being
retrieved. However, in other cases it may result in different data. Impromptu will always try to use the shortest
route in joining multiple tables. It will also try to use the tables that are already included in the query, rather
than including an additional table.
There is no hard and fast rule to resolving Loop Joins. There are four basic resolutions:
1. Break the join
2. Create alias tables with different join strategies
3. Use the join expression editor to specify the join
4. Modify SQL
Each of these resolutions is done for a different reason and may have some issues associated with it.
Determine the best resolution for your situation by analyzing the data with regards to the results required from
the join structure.

Example:
The join structure looks like this:
A=B
A=C
B=C
This is producing incorrect results. To resolve this issue, make table C an alias to omit the loop in the join
structure and this will result in data displaying correctly.
Correct Join Structure:
A=B
A = C alias
B=C
8.Use this query i am retriving all years OCt data from 01-10-2004 to 30-10-2007
i need to restrict this query to current date and current year
You have a function called 'extract' in cognos.
Ex:- extract(month, the date field). By giving like this you will get month. So you can keep a filter to restrict
the rows only for October.
9.How to show the data reported horizontally:
(For example:)
employee skill
1a
1b
1c
2d
2e
2f
Report result:
1 abc
2 def
Assuming 3 records per grouped item:
1. Group on employee
2. Create a running count based on the skill field.
3. Create 3 calculated columns based on the count field.
Call them skill1, skill2, skill3:
if (count = 1) then (skill) else null
if (count = 2) then (skill) else null
if (count = 3) then (skill) else null
4. Create 3 more calculated columns using the maximum function. Call them maxskill1, maxskill2, maxskill3
maximum (skill1)
maximum (skill2)
maximum (skill3)
5. Group on employee on maxskill1 on maxskill2 on maxskill3
6. Report employee axskill1 maxskill2 maxskill3
10.How to pass multiple values from picklist prompt to sub report filter
The sub-report only includes the first value.
When the sub-report query runs, it checks for the first row in the Customer Name column and shows only
information for that customer. If you want a sub-report to show information for another row in the column,

place the main report in a form frame that shows only one row at a time. When you insert the sub-report into
the form frame as well, it changes as you click through the rows in the main report. For example, the main
and sub-report above are both in a form frame that shows only one row of the Customer Name column at a
time. Each time you scroll to another customer name, the sub-report shows only information for that
customer."
11.How can I create a dynamic column name in Cognos
1. Create a calculated column which contains the information that the header is to contain, such as "Report
for year 1999" (concatenated text and date to string sub string extraction).
2. Highlight the report, and then right-click.
3. Select Properties, and then click the Headers/Footers tab.
4.Clear the Column Title Header check box. This will remove the headers from your columns.
5. Reinsert the rest of the column headers; insert text will work. For the dynamic column, from the Insert
menu, click Data and select the calculated column you created and insert it into the report.

Data warehouse Concepts


1.Why fact table is in normal form?
A fact table consists of measurements of business requirements and foreign keys of dimensions tables as
per business rules.
2.What is Difference between E-R Modeling and Dimentional Modeling.
Basic diff is E-R modeling will have logical and physical model. Dimensional model will have only physical
model.
E-R modeling is used for normalizing the OLTP database design.
Dimensional modeling is used for de-normalizing the ROLAP/MOLAP design.
3.What is conformed fact?
Conformed dimensions are the dimensions which can be used across multiple Data Marts in combination
with multiple facts tables accordingly
4.What are the methodologies of Data Warehousing.
Most of the time, we use Mr. Ralph Kimball methodologies for datawarehousing design.Two kind of schema
star and snow flake.
5.What is Data warehosuing Hierarchy?
Hierarchies
Hierarchies are logical structures that use ordered levels as a means of organizing data. A hierarchy can be
used to define data aggregation. For example, in a time dimension, a hierarchy might aggregate data from
the month level to the quarter level to the year level. A hierarchy can also be used to define a navigational
drill path and to establish a family structure.
Within a hierarchy, each level is logically connected to the levels above and below it. Data values at lower
levels aggregate into the data values at higher levels. A dimension can be composed of more than one
hierarchy. For example, in the product dimension, there might be two hierarchies--one for product categories
and one for product suppliers.
Dimension hierarchies also group levels from general to granular. Query tools use hierarchies to enable you
to drill down into your data to view different levels of granularity. This is one of the key benefits of a data
warehouse.
When designing hierarchies, you must consider the relationships in business structures. For example, a
divisional multilevel sales organization.
Hierarchies impose a family structure on dimension values. For a particular level value, a value at the next
higher level is its parent, and values at the next lower level are its children. These familial relationships
enable analysts to access data quickly.
Levels
A level represents a position in a hierarchy. For example, a time dimension might have a hierarchy that
represents data at the month, quarter, and year levels. Levels range from general to specific, with the root
level as the highest or most general level. The levels in a dimension are organized into one or more
hierarchies.
Level Relationships
Level relationships specify top-to-bottom ordering of levels from most general (the root) to most specific
information. They define the parent-child relationship between the levels in a hierarchy.
Hierarchies are also essential components in enabling more complex rewrites. For example, the database
can aggregate an existing sales revenue on a quarterly base to a yearly aggregation when the dimensional
dependencies between quarter and year are known.
6.What is surrogate key ? where we use it expalin with examples
Surrogate key is a substitution for the natural primary key.It is just a unique identifier or number for each row
that can be used for the primary key to the table. The only requirement for a surrogate primary key is that it
is unique for each row in the table.
Data warehouses typically use a surrogate, (also known as artificial or identity key), key for the dimension
tables primary keys. They can use Infa sequence generator, or Oracle sequence, or SQL Server Identity
values for the surrogate key.
It is useful because the natural primary key (i.e. Customer Number in Customer table) can change and this
makes updates more difficult.

Some tables have columns such as AIRPORT_NAME or CITY_NAME which are stated as the primary keys
(according to the business users) but ,not only can these change, indexing on a numerical value is probably
better and you could consider creating a surrogate key called, say, AIRPORT_ID. This would be internal to
the system and as far as the client is concerned you may display only the AIRPORT_NAME.
7.Importance of Surrogate Key in Data warehousing?
Surrogate Key is a Primary Key for a Dimension table. Most importance of using it is it is independent of
underlying database. i.e Surrogate Key is not affected by the changes going on with a database.
8.What is the flow of loading data into fact & dimensional tables?
Fact table - Table with Collection of Foreign Keys corresponding to the Primary Keys in Dimensional table.
Consists of fields with numeric values.
Dimension table - Table with Unique Primary Key.
Load - Data should be first loaded into dimensional table. Based on the primary key values in dimensional
table, the data should be loaded into Fact table.
9.What is a linked cube?
A cube can be stored on a single analysis server and then defined as a linked cube on other Analysis
servers. End users connected to any of these analysis servers can then access the cube. This arrangement
avoids the more costly alternative of storing and maintaining copies of a cube on multiple analysis servers.
linked cubes can be connected using TCP/IP or HTTP. To end users a linked cube looks like a regular cube.
10.What is meant by metadata in context of a Datawarehouse and how it is important?
Metadata or Meta Data Metadata is data about data. Examples of metadata include data element
descriptions, data type descriptions, attribute/property descriptions, range/domain descriptions, and
process/method descriptions. The repository environment encompasses all corporate metadata resources:
database catalogs, data dictionaries, and navigation services. Metadata includes things like the name,
length, valid values, and description of a data element. Metadata is stored in a data dictionary and
repository. It insulates the data warehouse from changes in the schema of operational systems. Metadata
Synchronization The process of consolidating, relating and synchronizing data elements with the same or
similar meaning from different systems. Metadata synchronization joins these differing elements together in
the data warehouse to allow for easier access.
11.Differentiate Primary Key and Partition Key?
Primary Key is a combination of unique and not null. It can be a collection of key values called as composite
primary key. Partition Key is a just a part of Primary Key. There are several methods of partition like Hash,
DB2, Random etc..While using Hash partition we specify the Partition Key.
12.What are the possible data marts in Retail sales.?
Product information, sales information
13.What is degenerate dimension table?
In simple terms, the column in a fact table that does not map to any dimensions, neither it s a measure
column. for e.g Invoice no, Invoice_line_no in fact table will be a degenerate dimension (columns), provided
if you dont have a dimension called invoice.
14.What is the main differnce between schema in RDBMS and schemas in DataWarehouse?
RDBMS Schema
* Used for OLTP systems
* Traditional and old schema
* Normalized
* Difficult to understand and navigate
* Cannot solve extract and complex problems
* Poorly modelled
DWH Schema
* Used for OLAP systems
* New generation schema
* De Normalized
* Easy to understand and navigate

* Extract and complex problems can be easily solved


* Very good model
15.Difference between Snow flake and Star Schema. What are situations where Snow flake Schema
is better than Star Schema to use and when the opposite is true?
Star schema contains the dimesion tables mapped around one or more fact tables.It is a denormalised
model.No need to use complicated joins.Queries results fastly.
Snowflake schema It is the normalised form of Star schema.contains indepth joins ,bcas the tbales r
splitted in to many pieces.We can easily do modification directly in the tables.We hav to use comlicated joins
,since we hav more tables .There will be some delay in processing the Query .
16. What is a CUBE in datawarehousing concept?
Cubes r muti-dimensional view of dw or data marts. it is designed in a logical way to drill, slice-n-dice. Every
part of the cube is a logical representation of the combination of facts-dimension attribs.
17. Is it correct/feasible develop a Data Mart using an ODS?
The ODS is technically designed to be used as the feeder for the DW and other DM's -- yes. It is to be the
source of truth.
18.Why should you put your data warehouse on a different system than your OLTP system
OLTP system stands for on-line transaction processing. These are used to store only daily transactions as
the changes have to be made in as few places as possible. OLTP do not have historical data of the
organization Datawarehouse will contain the historical information about the organization
19.Why are OLTP database designs not generally a good idea for a Data Warehouse
OLTP cannot store historical information about the organization. It is used for storing the details of daily
transactions while a datawarehouse is a huge storage of historical information obtained from different
datamarts for making intelligent decisions about the organization.
20.What are conformed dimensions
Conformed Dimensions are the Dimensions which are common to two cubes .say CUBE-1 contains
F1,D1,D2,D3 and CUBE-2 contains F2,D1,D2,D4 are the Facts and Dimensions ,here D1,D2 are the
Conformed Dimensions
21.What are Semi-additive and factless facts and in which scenario will you use such kinds of fact
tables
A fact less fact table captures the many-to-many relationships between dimensions, but contains no numeric
or textual facts. They are often used to record events or coverage information. Common examples of fact
less fact tables include:
- Identifying product promotion events (to determine promoted products that didnt sell)
- Tracking student attendance or registration events
- Tracking insurance-related accident events
- Identifying building, facility, and equipment schedules for a hospital or university
22.What is SCD1 , SCD2 , SCD3
The value of dimensions is used change very rarely, That is called Slowly Changing dimensions
Here mainly 3
1) SCD1: Replace the old values overwrite by new values
2) SCD2: Just Creating Additional records
3) SCD3: Its maintain just previous and recent
In the SCD2 again 3
1) Versioning
2) Flag value
3) Effective Date range
Versioning: Here the updated dimensions inserted in to the target along with version number
The new dimensions will be inserted into the target along with Primary key
Flagvalue:The updated dimensions insert into the target along with 0
and new dimensions inset into the target along with 1
OR
SCD Type 1, the attribute value is overwritten with the new value, obliterating the historical attribute values.
For example, when the product roll-up changes for a given product, the roll-up attribute is merely updated
with the current value.

SCD Type 2,a new record with the new attributes is added to the dimension table. Historical fact table rows
continue to reference the old dimension key with the old roll-up attribute; going forward, the fact table rows
will reference the new surrogate key with the new roll-up thereby perfectly partitioning history.
SCDType 3, attributes are added to the dimension table to support two simultaneous roll-ups - perhaps the
current product roll-up as well as current version minus one, or current version and original.
23.What is VLDB
The perception of what constitutes a VLDB continues to grow. A one terabyte database would normally be
considered to be a VLDB.
24.What are non-additive facts
Non-additive facts are facts that cannot be summed up for any of the dimensions present in the fact table.
Example: temparature,bill number...etc
25.What are slowly changing dimensions
If the data in the dimension table happen to change very rarely,then it is called as slowly changing
dimension.
Ex: changing the name and address of a person, this happens rarely.
26.What does level of Granularity of a fact table signify
In simple terms, level of granularity defines the extent of detail. As an example, let us look at geographical
level of granularity. We may analyze data at the levels of COUNTRY, REGION, TERRITORY, CITY and
STREET. In this case, we say the highest level of granularity is STREET.
27.Which columns go to the fact table and which columns go the dimension table
The Aggregation or calculated value columns will go to Fact Table and details information will go to
dimensional table.
28.What is ODS?
ODS means Operational Data Store It is used to store current data through transactional web applications,
Sap, MQ series Current data means particular data from one date into one date. ODS contains 30-90 data.
29.What is Normalization, First Normal Form, Second Normal Form , Third Normal Form
Normalization: The process of decomposing tables to eliminate data redundancy is called Normalization.
1 N.F :- The table should contain scalar or atomic values.
2 N.F :- Table should be in 1N.F + No partial functional dependencies
3 N.F :-Table should be in 2 N.F + No transitive dependencies
30.What is real time data-warehousing?
Real-time data warehousing is a combination of two things: 1) real-time activity and 2) data warehousing.
Real-time activity is activity that is happening right now. The activity could be anything such as the sale of
widgets. Once the activity is complete, there is data about it. Data warehousing captures business activity
data. Real-time data warehousing captures business activity data as it occurs. As soon as the business
activity is complete and there is data about it, the completed activity data flows into the data warehouse and
becomes available instantly. In other words, real-time data warehousing is a framework for deriving
information from data as the data becomes available.
31.What are modeling tools available in the Market
Modeling Tool Vendor
Erwin Computer Associates
ER/Studio Embarcadero
Power Designer Sybase
Oracle Designer Oracle
32.What is a general purpose scheduling tool?
General purpose of scheduling tool may be cleansing and loading data at specific given time.
33.What is a lookup table?
reference table can be otherwise called as lookup table

34.What is a dimension table?


A dimensional table is a collection of hierarchies and categories along which the user can drill down and drill
up. it contains only the textual attributes.
35.What is Fact table?
A table in a data warehouse whose entries describe data in a fact table. Dimension tables contain the data
from which dimensions are created.
36.What are the various Reporting tools in the Market?
INEA
MS-Excel
Business Objects (Crystal Reports)
Cognos (Impromptu, Power Play)
Micro strategy
MS reporting services
Informatic Power Analyzer
Actuate
Hyperion (BRIO)
Oracle Express OLAP
Proclarity
SAS
37.What are the vaious ETL tools in the Market?
1. Informatica Power Center
2. Ascential Data Stage
3. ESS Base Hyperion
4. AbIntio
5. BO Data Integrator
6. SAS ETL
7. MS DTS
8. Oracle OWB
9. Pervasive Data Junction
10. Cognos Decision Stream
38.What is ETL
ETL is an abbreviation for "Extract, Transform and Load. This is the process of extracting data from their
operational data sources or external data sources, transforming the data which includes cleansing,
aggregation, summarization, integration, as well as basic transformation and loading the data into some
form of the data warehouse.
39.What is the Difference between OLTP and OLAP
Current data
Short database transactions
Online update/insert/delete
Normalization is promoted
High volume transactions
Transaction recovery is necessary
OLAP
Current and historical data
Long database transactions
Batch update/insert/delete
Denormalization is promoted
Low volume transactions
Transaction recovery is not necessary
40.What are Aggregate tables?
Aggregate table contains the summary of existing warehouse data which is grouped to certain levels of
dimensions.Retrieving the required data from the actual table, which have millions of records will take more
time and also affects the server performance.To avoid this we can aggregate the table to certain required
level and can use it.This tables reduces the load in the database server and increases the performance of
the query and can retrieve the result very fastly. E.g Yearly, monthly sales information.

41.What are the Different methods of loading Dimension tables?


Conventional Load:
Before loading the data, all the Table constraints will be checked against the data.
Direct load:(Faster Loading)
All the Constraints will be disabled. Data will be loaded directly. Later the data will be checked against the
table constraints and the bad data won't be indexed.
42.What is Dimensional Modeling?
Dimensional Modeling is a design concept used by many data warehouse designers to build their data
warehouse. In this design model all the data is stored in two types of tables - Facts table and Dimension
table. Fact table contains the facts/measurements of the business and the dimension table contains the
context of measurements i.e., the dimensions on which the facts are calculated.
43.What is ER Diagram
ER - Stands for entitity relationship diagrams. It is the first step in the design of data model which will later
lead to a physical database design of possible a OLTP or OLAP database.
44.What are Data Marts
Data Mart is a segment of a data warehouse that can provide data for reporting and analysis on a section,
unit, department or operation in the company, e.g. sales, payroll, production. Data marts are sometimes
complete individual data warehouses which are usually smaller than the corporate data warehouse.
45.What is a Data Warehousing?
Data Warehouse is a repository of integrated information, available for queries and analysis. Data and
information are extracted from heterogeneous sources as they are generated....This makes it much easier
and more efficient to run queries over data that originally came from different sources. Typical relational
databases are designed for on-line transactional processing (OLTP) and do not meet the requirements for
effective on-line analytical processing (OLAP). As a result, data warehouses are designed differently than
traditional relational databases.

Informatica
1.What are Target Options on the Servers?
Target Options for File Target type are FTP File, Loader and MQ. There are no target options for ERP target
type Target Options for Relational are Insert, Update (as Update), Update (as Insert), Update (else Insert),
Delete, and Truncate Table
2.How do you identify existing rows of data in the target table using lookup transformation?
Can identify existing rows of data using unconnected lookup transformation.
3.What are Aggregate transformation?
The Aggregator transformation allows you to perform aggregate calculations, such as averages and sums.
The Aggregator transformation is unlike the Expression transformation, in that you can use the Aggregator
transformation to perform calculations on groups
4.What are various types of Aggregation?
Various types of aggregation are SUM, AVG, COUNT, MAX, MIN, FIRST, LAST, MEDIAN, PERCENTILE,
STDDEV, and VARIANCE.
5.What are 2 modes of data movement in Informatica Server?
The data movement mode depends on whether Informatica Server should process single byte or multi-byte
character data. This mode selection can affect the enforcement of code page relationships and code page
validation in the Informatica Client and Server.
a) Unicode - IS allows 2 bytes for each character and uses additional byte for each nonascii character (such as Japanese characters)
b) ASCII - IS holds all data in a single byte
The IS data movement mode can be changed in the Informatica Server configuration parameters. This
comes into effect once you restart the Informatica Server.
6..What is Code Page Compatibility?
Compatibility between code pages is used for accurate data movement when the Informatica Sever runs in
the Unicode data movement mode. If the code pages are identical, then there will not be any data loss. One
code page can be a subset or superset of another. For accurate data movement, the target code page must
be a superset of the source code page.
Superset - A code page is a superset of another code page when it contains the character encoded in the
other code page, it also contains additional characters not contained in the other code page.
Subset - A code page is a subset of another code page when all characters in the code page are encoded in
the other code page.
7.What is Code Page used for?
Code Page is used to identify characters that might be in different languages. If you are importing Japanese
data into mapping, u must select the Japanese code page of source data.
8.What is Router transformation?
Router transformation allows you to use a condition to test data. It is similar to filter transformation. It allows
the testing to be done on one or more conditions. Router transformation is use to load data in multiple
targets depending on the test condition.
9.What is Load Manager?
While running a Workflow,the PowerCenter Server uses the Load Manager process and the Data
Transformation Manager Process (DTM) to run the workflow and carry out workflow tasks.When the
PowerCenter Server runs a workflow, the Load Manager performs the following tasks:
1. Locks the workflow and reads workflow properties.
2. Reads the parameter file and expands workflow variables.
3. Creates the workflow log file.
4. Runs workflow tasks.
5. Distributes sessions to worker servers.
6. Starts the DTM to run sessions.
7. Runs sessions from master servers.
8. Sends post-session email if the DTM terminates abnormally.

When the PowerCenter Server runs a session, the DTM performs the following tasks:
1. Fetches session and mapping metadata from the repository.
2. Creates and expands session variables.
3. Creates the session log file.
4. Validates session code pages if data code page validation is enabled. Checks query
conversions if data code page validation is disabled.
5. Verifies connection object permissions.
6. Runs pre-session shell commands.
7. Runs pre-session stored procedures and SQL.
8. Creates and runs mapping, reader, writer, and transformation threads to extract,transform, and load data.
9. Runs post-session stored procedures and SQL.
10. Runs post-session shell commands.
11. Sends post-session email.
10.What is Data Transformation Manager?
After the load manager performs validations for the session, it creates the DTM process. The DTM process
is the second process associated with the session run. The primary purpose of the DTM process is to create
and manage threads that carry out the session tasks.
The DTM allocates process memory for the session and divide it into buffers. This is also
known as buffer memory. It creates the main thread, which is called the master thread.
The master thread creates and manages all other threads.
If we partition a session, the DTM creates a set of threads for each partition to allow
concurrent processing.. When Informatica server writes messages to the session log it
includes thread type and thread ID. Following are the types of threads that DTM creates:
Master thread - Main thread of the DTM process. Creates and manages all other threads. Mapping thread One Thread to Each Session. Fetches Session and Mapping Information. Pre and Post Session Thread-One
Thread each to Perform Pre and Post Session Operations.reader thread-One Thread for Each Partition for
Each Source Pipeline. WRITER THREAD-One Thread for Each Partition if target exist in the source pipeline
write to the target. TRANSFORMATION THREAD - One or More Transformation Thread For Each Partition
11.What is Session and Batches?
Session - A Session Is A set of instructions that tell the Informatica Server How and When to Move Data
from Sources to Targets. After creating the session, we can use either the server manager or the command
line program pmcmd to start or stop the session.Batches - It Provides A Way to Group Sessions for Either
Serial or Parallel Execution by the Informatica Server. There Are Two Types Of Batches :
Sequential - Run Session One after the Other. concurrent - Run Session At The Same Time.
12.What is a source qualifier?
When you add a relational or a flat file source definition to a mapping, you need to connect it to a Source
Qualifier transformation. The Source Qualifier represents the rows that the Informatica Server reads when it
executes a session
13.Why we use lookup transformations?
Lookup Transformations can access data from relational tables that are not sources in mapping. With
Lookup transformation, we can accomplish the following tasks:
Get a related value-Get the Employee Name from Employee table based on the Employee ID
Perform Calculation.
Update slowly changing dimension tables - We can use unconnected lookup transformation to
determine whether the records already exist in the target or not.
14.While importing the relational source defintion from database,what are the meta data of source U
import?
Source name
Database location
Column names
Data types
Key constraints

15.How many ways you can update a relational source defintion and what r they?
Two ways
1. Edit the definition
2. Reimport the definition
16.Where should U place the flat file to import the flat file defintion to the designer?
There is no such restriction to place the source file. In performance point of view its better to place the file in
server local source folder. if you need path please check the server properties available at workflow
manager. It doesn't mean we should not place in any other folder, if we place in server source folder by
default source will be selected at time session creation.
17.To provide support for Mainframes source data,which files r used as a source definitions?
COBOL Copy-book files
18.Which transformation should u need while using the cobol sources as source defintions?
Normalizer transformation which is used to normalize the data. Since cobol sources r oftenly consists of
Demoralized data.
19.How can U create or import flat file definition in to the warehouse designer?
U can not create or import flat file defintion in to warehouse designer directly. Instead U must analyze the file
in source analyzer, then drag it into the warehouse designer. When U drag the flat file source definition into
warehouse designer workspace, the warehouse designer creates a relational target defintion not a file
defintion.If u want to load to a file, configure the session to write to a flat file. When the informatica server
runs the session, it creates and loads the flat file.
20.What is the maplet?
Maplet is a set of transformations that you build in the maplet designer and U can use in multiple mapings.
For Ex:Suppose we have several fact tables that require a series of dimension keys.Then we can create a
mapplet which contains a series of Lkp transformations to find each dimension key and use it in each fact
table mapping instead of creating the same Lkp logic in each mapping. Set of transforamations where the
logic can be reusable.
21.What is a transforamation?
A transformation is repository object that pass data to the next stage (i.e to the next transformation or target)
with/with out modifying the data
22.What r the designer tools for creating tranformations?
Mapping designer
Transformation developer
Mapplet designer
23.What r the active and passive transforamtions?
Transformations can be active or passive. An active transformation can change the number of rows that
pass through it, such as a Filter transformation that removes rows that do not meet the filter condition.
A passive transformation does not change the number of rows that pass through it, such as an Expression
transformation that performs a calculation on data and passes all rows through the transformation.
24.What r the connected or unconnected transforamations?
An unconnected transforamtion is not connected to other transformations in the mapping.Connected
transforamation is connected to other transforamtions in the mapping.
25.How many ways u create ports?
Two ways
1.Drag the port from another transformation
2.Click the add button on the ports tab.

26.What r the reusable transforamtions?


Reusable transformations can be used in multiple mappings. When u need to incorporate this transformation
into mapping add an instance of it to maping.Later if U change the definition of the transformation, all
instances of it inherit the changes. Since the instance of reusable transformation is a pointer to that
transformation can change the transformation in the transformation developer, its instances automatically
reflect these changes this feature can save U great deal of work.
27.What r the methods for creating reusable transforamtions?
Two methods
1.Design it in the transformation developer.
2. Promote a standard transformation from the mapping designer. After U adds a transformation to the
mapping, U can promote it to the status of reusable transformation. Once U promotes a standard
transformation to reusable status can demote it to a standard transformation at any time.
If u change the properties of a reusable transformation in mapping can revert it to the original reusable
transformation properties by clicking the revert button.
28.What r the unsupported repository objects for a mapplet?
COBOL source definition
Joiner transformations
Normalizer transformations
Non reusable sequence generator transformations.
Pre or post session stored procedures
Target defintions
Power mart 3.5 style Look Up functions
XML source definitions
IBM MQ source definitions
29. What r the mapping paramaters and maping variables?
Mapping parameter represents a constant value that U can define before running a session. A mapping
parameter retains the same value throughout the entire session. When u use the mapping parameter, U
declare and use the parameter in a mapping or maplet.Then define the value of parameter in a parameter
file for the session. Unlike a mapping parameter, a mapping variable represents a value that can change
throughout the session. The informatica server saves the value of mapping variable to the repository at the
end of session run and uses that value next time U run the session.
30.Can U use the maping parameters or variables created in one maping into another maping?
N0.We can use mapping parameters or variables in any transformation of the same maping or mapplet in
which U have created maping parameters or variables.
31. Can u use the maping parameters or variables created in one maping into any other reusable
transformation?
Yes.Because reusable tranformation is not contained with any maplet or maping.
32. How can U improve session performance in aggregator transformation?
Use sorted input.
33.What is aggregate cache in aggregator transforamtion?
The aggregator stores data in the aggregate cache until it completes aggregate calculations.When u run a
session that uses an aggregator transformation,the informatica server creates index and data caches in
memory to process the transformation.If the informatica server requires more space,it stores overflow values
in cache files.
34.What r the diffrence between joiner transformation and source qualifier transformation?
U can join hetrogenious data sources in joiner transformation which we can not achieve in source qualifier
transformation. U need matching keys to join two relational sources in source qualifier transformation.Where
as u doesnt need matching keys to join two sources. Two relational sources should come from same
datasource in sourcequalifier.U can join relatinal sources which r coming from diffrent sources also.

35. In which condtions we can not use joiner transformation(Limitaions of joiner transformation)?
Both pipelines begin with the same original data source.
Both input pipelines originate from the same Source Qualifier transformation.
Both input pipelines originate from the same Normalizer transformation.
Both input pipelines originate from the same Joiner transformation.
Either input pipelines contains an Update Strategy transformation.
Either input pipelines contains a connected or unconnected Sequence Generator transformation.
36. what r the settiings that u use to cofigure the joiner transformation?
Master and detail source
Type of join
Condition of the join
37. What r the join types in joiner transformation?
Normal (Default)
Master outer
Detail outer
Full outer
38.What r the joiner caches?
When a Joiner transformation occurs in a session, the Informatica Server reads all the records from the
master source and builds index and data caches based on the master rows. After building the caches, the
Joiner transformation reads records from the detail source and perform joins.
39.what is the look up transformation?
Use lookup transformation in ur mapping to lookup data in a relational table,view,synonym.
Informatica server queries the look up table based on the lookup ports in the transformation.It compares the
lookup transformation port values to lookup table column values based on the look up condition.
40. Why use the lookup transformation ?
To perform the following tasks.
Get a related value. For example, if your source table includes employee ID, but you want to include the
employee name in your target table to make your summary data easier to read. Perform a calculation. Many
normalized tables include values used in a calculation, such as gross sales per invoice or sales tax, but not
the calculated value (such as net sales). Update slowly changing dimension tables. You can use a Lookup
transformation to determine whether records already exist in the target.
41. What r the types of lookup?
Connected and unconnected.
42. Differences between connected and unconnected lookup?
Connected lookup
Unconnected lookup
Receives input values diectly from the Receives input values from the result of a lkp expression in
pipe line.
a another transformation.
U can use a dynamic or static cache U can use a static cache.
Cache includes all lookup columns
Cache includes all lookup out put ports in the lookup
used in the maping
condition and the lookup/return port.
Support user defined default values
Does not support user defiend default
43.What is meant by lookup caches?
The informatica server builds a cache in memory when it processes the first row af a data in a cached look
up transformation.It allocates memory for the cache based on the amount u configure in the transformation
or session properties.The informatica server stores condition values in the index cache and output values in
the data cache.
44. What r the types of lookup caches?

Persistent cache: U can save the lookup cache files and reuse them the next time the informatica server
processes a lookup transformation configured to use the cache.
Recache from database: If the persistent cache is not synchronized with he lookup table,U can configure the
lookup transformation to rebuild the lookup cache.
Static cache: U can configure a static or readonly cache for only lookup table.By default informatica server
creates a static cache.It caches the lookup table and lookup values in the cache for each row that comes
into the transformation.when the lookup condition is true,the informatica server does not update the cache
while it prosesses the lookup transformation.
Dynamic cache: If u want to cache the target table and insert new rows into cache and the target,u can
create a look up transformation to use dynamic cache.The informatica server dynamically inerts data to the
target table.
shared cache: U can share the lookup cache between multiple transactions.U can share unnamed cache
between transformations inthe same maping.
45.Difference between static cache and dynamic cache
Static cache

Dynamic cache
U can insert rows into the cache as u pass
U can not insert or update the cache
to the target
The informatic server returns a value from the lookup table or The informatic server inserts rows into
cache when the condition is true.When the condition is not
cache when the condition is false.This
true, informatica server returns the default value for
indicates that the the row is not in the cache
connected transformations and null for unconnected
or target table. U can pass these rows to the
transformations.
target table
46.Which transformation should we use to normalize the COBOL and relational sources?
Normalizer Transformation. When U drag the COBOL source in to the mapping Designer workspace, the
normalizer transformation automatically appears, creating input and output ports for every column in the
source.
47.How the informatica server sorts the string values in Ranktransformation?
When the informatica server runs in the ASCII data movement mode it sorts session data using Binary
sortorder.If U configure the seeion to use a binary sort order,the informatica server caluculates the binary
value of each string and returns the specified number of rows with the higest binary values for the string.
48.What r the rank caches?
During the session, the informatica server compares an inout row with rows in the datacache.If the input row
out-ranks a stored row, the informatica server replaces the stored row with the input row. The informatica
server stores group information in an index cache and row data in a data cache.
49. What is the Rankindex in Ranktransformation?
The Designer automatically creates a RANKINDEX port for each Rank transformation. The Informatica
Server uses the Rank Index port to store the ranking position for each record in a group. For example, if you
create a Rank transformation that ranks the top 5 salespersons for each quarter, the rank index numbers the
salespeople from 1 to 5:
50. What is the Router transformation?
A Router transformation is similar to a Filter transformation because both transformations allow you to use a
condition to test data. However, a Filter transformation tests data for one condition and drops the rows of
data that do not meet the condition. A Router transformation tests data for one or more conditions and gives
you the option to route rows of data that do not meet any of the conditions to a default output group.
If you need to test the same input data based on multiple conditions, use a Router Transformation in a
mapping instead of creating multiple Filter transformations to perform the same task.

51.What r the types of groups in Router transformation?


Input group Output group
The designer copies property information from the input ports of the input group to create a set of output
ports for each output group.
Two types of output groups
User defined groups
Default group
U can not modify or delete default groups.
52. Why we use stored procedure transformation?
For populating and maintaining data bases.
53. What r the types of data that passes between informatica server and stored procedure?
3 types of data
Input/Out put parameters
Return Values
Status code.
54. What is the status code?
Status code provides error handling for the informatica server during the session.The stored procedure
issues a status code that notifies whether or not stored procedure completed sucessfully.This value can not
seen by the user.It only used by the informatica server to determine whether to continue running the session
or stop.
55. What is source qualifier transformation?
When U add a relational or a flat file source definition to a mapping need to connect it to
a source qualifier transformation.The source qualifier transformation represents the records
that the informatica server reads when it runs a session
56.What r the tasks that source qualifier performs?
Join data originating from same source data base.
Filter records when the informatica server reads source data.
Specify an outer join rather than the default inner join
specify sorted records.
Select only distinct values from the source.
Creating custom query to issue a special SELECT statement for the informatica server to read
source data.
57.What is the target load order?
U specify the target loadorder based on source qualifiers in a maping.If u have the multiple
source qualifiers connected to the multiple targets,U can designatethe order in which informatica
server loads data into the targets.
58. What is the default join that source qualifier provides?
Inner equi join
59.What r the basic needs to join two sources in a source qualifier?
Two sources should have primary and Foreign key relation ships.
Two sources should have matching data types.
60.What is update strategy transformation ?
This transformation is used to maintain the history data or just most recent changes in to target
table.

61.Describe two levels in which update strategy transformation sets?


Within a session. When you configure a session, you can instruct the Informatica Server to either treat all
records in the same way (for example, treat all records as inserts), or use instructions coded into the session
mapping to flag records for different database operations. Within a mapping. Within a mapping, you use the
Update Strategy transformation to flag records for insert, delete, update, or reject.
62.What is the default source option for update stratgey transformation?
Data driven
63.What is Datadriven?
The informatica server follows instructions coded into update strategy transformations with in the session
maping determine how to flag records for insert, update, delete or reject. If u do not choose data driven
option setting,the informatica server ignores all update strategy transformations in the mapping.
64.What r the options in the target session of update strategy transsformatioin?
Insert
Delete
Update
Update as update
Update as insert
Update esle insert
Truncate table
65.What r the types of mapping wizards that r to be provided in Informatica?
The Designer provides two mapping wizards to help you create mappings quickly and easily. Both wizards
are designed to create mappings for loading and maintaining star schemas, a series of dimensions related
to a central fact table.
Getting Started Wizard. Creates mappings to load static fact and dimension tables, as well as slowly
growing dimension tables.
Slowly Changing Dimensions Wizard. Creates mappings to load slowly changing dimension tables based on
the amount of historical dimension data you want to keep and the method you choose to handle historical
dimension data.
66.What r the types of maping in Getting Started Wizard?
Simple Pass through mapping :
Loads a static fact or dimension table by inserting all rows. Use this mapping when you want to drop all
existing data from your table before loading new data.
Slowly Growing target:
Loads a slowly growing fact or dimension table by inserting new rows. Use this mapping to load new data
when existing data does not require updates.
67.What r the mapings that we use for slowly changing dimension table?
Type1: Rows containing changes to existing dimensions are updated in the target by overwriting the existing
dimension. In the Type 1 Dimension mapping, all rows contain current dimension data.
Use the Type 1 Dimension mapping to update a slowly changing dimension table when you do not need to
keep any previous versions of dimensions in the table.
Type 2: The Type 2 Dimension Data mapping inserts both new and changed dimensions into the target.
Changes are tracked in the target table by versioning the primary key and creating a version number for
each dimension in the table.
Use the Type 2 Dimension/Version Data mapping to update a slowly changing dimension table when you
want to keep a full history of dimension data in the table. Version numbers and versioned primary keys track
the order of changes to each dimension.
Type 3: The Type 3 Dimension mapping filters source rows based on user-defined comparisons and inserts
only those found to be new dimensions to the target. Rows containing changes to existing dimensions are
updated in the target. When updating an existing dimension, the Informatica Server saves existing data in
different columns of the same row and replaces the existing data with the updates

68.What r the different types of Type2 dimension maping?


Type2 Dimension/Version Data Maping: In this maping the updated dimension in the source will gets
inserted in target along with a new version number.And newly added dimension
in source will inserted into target with a primary key.
Type2 Dimension/Flag current Maping: This maping is also used for slowly changing dimensions.In addition
it creates a flag value for changed or new dimension.
Flag indiactes the dimension is new or newlyupdated.Recent dimensions will gets saved with cuurent flag
value 1. And updated dimensions r saved with the value 0.
Type2 Dimension/Effective Date Range Mapping: This is also one flavor of Type2 mapping used for slowly
changing dimensions. This mapping also inserts both new and changed dimensions in to the target. And
changes r tracked by the effective date range for each version of each dimension.
69.How can u recognize whether or not the newly added rows in the source r gets insert in the
target?
In the Type2 mapping we have three options to recognize the newly added rows
Version number
Flag value
Effective date Range
70.What r two types of processes that informatica runs the session?
Load manager Process: Starts the session, creates the DTM process, and sends post-session email when
the session completes.
The DTM process. Creates threads to initialize the session, read, write, and transform data, and handle preand post-session operations.
71.What r the new features of the server manager in the informatica 5.0?
U can use command line arguments for a session or batch. This allows U to change the values of session
parameters, and mapping parameters and mapping variables.
Parallel data processing: This feature is available for power center only. If we use the informatica server on a
SMP system can use multiple CPUs to process a session concurrently.
Process session data using threads: Informatica server runs the session in two processes. Explained in
previous question.
72.Can u generate reports in Informatcia?
Yes. By using Metadata reporter we can generate reports in informatica.
73.What is metadata reporter?
It is a web based application that enables you to run reports against repository metadata.
With a Meta data reporter can access information about Ur repository with out having knowledge of sql,
transformation language or underlying tables in the repository.
74. Define mapping and sessions?
Mapping: It is a set of source and target definitions linked by transformation objects that define the rules for
transformation.
Session: It is a set of instructions that describe how and when to move data from source to targets.
75.Which tool U use to create and manage sessions and batches and to monitor and stop the
informatica server?
Informatica server manager.
76.Why we use partitioning the session in informatica?
Performance can be improved by processing data in parallel in a single session by creating multiple
partitions of the pipeline.Informatica server can achieve high performance by partitioning the pipeline and
performing the extract , transformation, and load for each partition in parallel.
77.To achieve the session partition what r the necessary tasks u have to do?
Configure the session to partition source data. Install the informatica server on a machine with multiple
CPUs.

78.How the informatica server increases the session performance through partitioning the source?
For a relational sources informatica server creates multiple connections for each parttion of a single source
and extracts separate range of data for each connection.Informatica server reads multiple partitions of a
single source concurently.Similarly for loading also informatica server creates multiple connections to the
target and loads partitions of data concurently.
For XML and file sources,informatica server reads multiple files concurently.For loading the data informatica
server creates a separate file for each partition(of a source file).U can choose to merge the targets.
79.Why u use repository connectivity?
When u edit,schedule the sesion each time,informatica server directly communicates the repository to check
whether or not the session and users r valid.All the metadata of sessions and mappings will be stored in
repository.
80.What r the tasks that Loadmanger process will do?
Manages the session and batch scheduling: When u start the informatica server the load maneger launches
and queries the repository for a list of sessions configured to run on the informatica server.When u configure
the session the load manager maintains list of list of sessions and session start times. When u sart a
session load manger fetches the session information from the repository to perform the validations and
verifications prior to starting DTM process. Locking and reading the session: When the informatica server
starts a session load manager locks the session from the repository. Locking prevents U starting the session
again and again. Reading the parameter file: If the session uses a parameter files, load manager reads the
parameter file and verifies that the session level parameters are declared in the file Verifies permission and
privileges: When the sesson starts load manger checks whether or not the user have privileges to run the
session.
Creating log files: Load manger creates log file contains the status of session.
81.What is DTM process?
After the loadmanger performs validations for session,it creates the DTM process.DTM is to create and
manage the threads that carry out the session tasks.I creates the master thread.Master thread creates and
manges all the other threads.
82.What r the different threads in DTM process?
Master thread: Creates and manages all other threads
Mapping thread: One mapping thread will be creates for each session.Fectches session and mapping
information.
Pre and post session threads: This will be created to perform pre and post session operations.
Reader thread: One thread will be created for each partition of a source.It reads data from source.
Writer thread: It will be created to load data to the target.
83.What r the data movement modes in informatcia?
Data movement modes determines how informatcia server handles the character data.U choose the data
movement in the informatica server configuration settings. Two types of data movement modes available in
Informatica.
ASCIImode
Uni code mode.
84.What r the out put files that the informatica server creates during the session running?
Informatica server log: Informatica server(on unix) creates a log for all status and error messages(default
name: pm.server.log).It also creates an error log for error messages.These files will be created in informatica
home directory.
Session log file: Informatica server creates session log file for each session.It writes information about
session into log files such as initialization process, creation of sql commands for reader and writer threads,
errors encountered and load summary. The amount of detail in session log file depends on the tracing level
that u set.

Session detail file: This file contains load statistics for each targets in mapping. Session detail include
information such as table name, number of rows written or rejected. U can view this file by double clicking on
the session in monitor window
Performance detail file: This file contains information known as session performance details which helps U
where performance can be improved. To generate this file select the performance detail option in the
session property sheet.
Reject file: This file contains the rows of data that the writer does not write to targets.
Control file: Informatica server creates control file and a target file when U runs a session that uses the
external loader. The control file contains the information about the target flat file such as data format and
loading instructions for the external loader.
Post session email: Post session email allows U to automatically communicate information about a session
run to designated recipients can create two different messages. One if the session completed successfully
the other if the session fails.
Indicator file: If u use the flat file as a target can configure the informatica server to create indicator file. For
each target row, the indicator file contains a number to indicate whether the row was marked for insert,
update, delete or reject.
Output file: If session writes to a target file, the informatica server creates the target file based on file
properties entered in the session property sheet.
Cache files: When the informatica server creates memory cache it also creates cache files. For the
following circumstances informatica server creates index and data cache files.
Aggregator transformation
Joiner transformation
Rank transformation
Lookup transformation
85.In which circumstances that informatica server creates Reject files?
When it encounters the DD_ Reject in update strategy transformation. Violates database constraint Filed in
the rows was truncated or overflowed.
86.What is polling?
It displays the updated information about the session in the monitor window. The monitor window displays
the status of each session when U poll the informatica server
87.Can u copy the session to a different folder or repository?
Yes. By using copy session wizard u can copy a session in a different folder or repository. But that
target folder or repository should consists of mapping of that session. If target folder or repository is not
having the mapping of copying session, u should have to copy that mapping first before u copy the session
88.What is batch and describe about types of batches?
Grouping of session is known as batch. Batches r two types
Sequential: Runs sessions one after the other
Concurrent: Runs session at same time.
If u has sessions with source-target dependencies u have to go for sequential batch to start the
sessions one after another. If u have several independent sessions u can use concurrent batches.
Which runs all the sessions at the same time.
89.Can u copy the batches?
NO
90.How many number of sessions that u can create in a batch?
Any number of sessions.
91.When the informatica server marks that a batch is failed?
If one of session is configured to "run if previous completes" and that previous session fails.

92.What is a command that used to run a batch?


pmcmd is used to start a batch.
93.What r the different options used to configure the sequential batches?
Two options
Run the session only if previous session completes successfully. Always runs the session.
94.In a sequential batch can u run the session if previous session fails?
Yes.By setting the option always runs the session.
95.Can u start a batches with in a batch?
U can not. If u wants to start batch that resides in a batch, create a new independent batch and copy the
necessary sessions into the new batch.
96.Can u start a session inside a batch individually?
We can start our required session only in case of sequential batch. in case of concurrent batch
we cant do like this.
97.How can u stop a batch?
By using server manager or pmcmd.
98.What r the session parameters?
Session parameters r like mapping parameters, represent values U might want to change between
sessions such as database connections or source files. Server manager also allows U to create user defined
session parameters. Following r user defined session parameters.
Database connections Source file names: use this parameter when u wants to change the name or
location of session source file between session runs
Target file name: Use this parameter when u want to change the name or location of session target file
between session runs.
Reject file name: Use this parameter when u want to change the name or location of
session reject files between session runs.
99.What is parameter file?
Parameter file is to define the values for parameters and variables used in a session. A parameter
file is a file created by text editor such as word pad or notepad. U can define the following values in
parameter file
Mapping parameters
Mapping variables
session parameters
100.How can u access the remote source into Ur session?
Relational source: To access relational source which is situated in a remote place, u need to
configure database connection to the data source.
File Source: To access the remote source file U must configure the FTP connection to the
host machine before u create the session.
Heterogeneous: When Ur mapping contains more than one source type, the server manager creates
a heterogeneous session that displays source options for all types.
101.What r the transformations that restricts the partitioning of sessions?
Advanced External procedure transformation and External procedure transformation: This
transformation contains a check box on the properties tab to allow partitioning.
Aggregator Transformation: If u use sorted ports u can not partition the associated source
Joiner Transformation: U can not partition the master source for a joiner transformation
Normalizer Transformation
XML targets.

102.Performance tuning in Informatica?


The goal of performance tuning is optimize session performance so sessions run during the available load
window for the Informatica Server. Increase the session performance by following.
The performance of the Informatica Server is related to network connections. Data generally moves across a
network at less than 1 MB per second, whereas a local disk moves data five to twenty times faster. Thus
network connections ofteny affect on session performance. So aviod netwrok connections.
Flat files: If ur flat files stored on a machine other than the informatca server, move those files to the
machine that consists of informatica server. Relational data sources: Minimize the connections to sources,
targets and informatica server to improve session performance. Moving target database into server system
may improve session performance. Staging areas: If u use staging areas u force informatica server to
perform multiple data passes. Removing of staging areas may improve session performance. U can run the
multiple informatica servers against the same repository.Distibuting the session load to multiple informatica
servers may improve session performance. Run the informatica server in ASCII data movement mode
improves the session performance. Because ASCII data movement mode stores a character value in one
byte. Unicode mode takes 2 bytes to store a character. If a session joins multiple source tables in one
Source Qualifier, optimizing the query may improve performance. Also, single table select statements with
an ORDER BY or GROUP BY clause may benefit from optimization such as adding indexes.
We can improve the session performance by configuring the network packet size, which allows
data to cross the network at one time.To do this go to server manger ,choose server configure database
connections. If u r target consists key constraints and indexes u slow the loading of data. To improve the
session performance in this case drop constraints and indexes before u run the session and rebuild them
after completion of session. Running parallel sessions by using concurrent batches will also reduce the time
of loading the data. So concurrent batches may also increase the session performance.
Partitioning the session improves the session performance by creating multiple connections to sources and
targets and loads data in paralel pipe lines. In some cases if a session contains a aggregator
transformation ,u can use incremental aggregation to improve session performance. Avoid transformation
errors to improve the session performance. If the sessioin contained lookup transformation u can improve
the session performance by enabling the look up cache. If Ur session contains filter transformation, create
that filter transformation nearer to the sources or u can use filter condition in source qualifier. Aggregator,
Rank and joiner transformation may often decrease the session performance .Because they must group
data before processing it. To improve session performance in this case use sorted ports option.
103.What is difference between maplet and reusable transformation?
Mapplet consists of set of transformations that is reusable. A reusable transformation is a
single transformation that can be reusable. If u creates a variables or parameters in maplet that can not be
used in another mapping or maplet.Unlike the variables that r created in a reusable transformation can be
useful in any other maping or maplet. We can not include source definitions in reusable transformations.But
we can add sources to a maplet. Whole transformation logic will be hided in case of maplet.But it is
transparent in case of reusable transformation. We cant use COBOL source qualifier, joiner, normalizer
transformations in mapplet. Where as we can make them as a reusable transformations.
104.Define informatica repository?
The Informatica repository is a relational database that stores information, or metadata, used by the
Informatica Server and Client tools. Metadata can include information such as mappings describing how to
transform source data, sessions indicating when you want the Informatica Server to perform the
transformations, and connect strings for sources and targets. The repository also stores administrative
information such as usernames and passwords, permissions and privileges, and product version. Use
repository manager to create the repository. The Repository Manager connects to the repository database
and runs the code needed to create the repository tables. These tables stores metadata in specific format
the informatica server, client tools use.
105.What r the types of metadata that stores in repository?
Following r the types of metadata that stores in the repository
Database connections
Global objects
Mappings
Mapplets
Multidimensional metadata
Reusable transformations
Sessions and batches
Short cuts

Source definitions
Target definitions
Transformations
106.What is power center repository?
The Power Center repository allows you to share metadata across repositories to create a data mart
domain. In a data mart domain, you can create a single global repository to store metadata used across an
enterprise, and a number of local repositories to share the global metadata as needed.
107.What r the new features in Informatica 5.0?
U can Debug Ur mapping in mapping designer
U can view the work space over the entire screen
The designer displays a new icon for a invalid mappings in the navigator window
U can use a dynamic lookup cache in a lookup transformation
Create mapping parameters or mapping variables in a mapping or maplet to make mappings more
flexible
U can export objects into repository and import objects from repository. when u export a repository object,
the designer or server manager creates an XML file to describe the repository metadata.
The designer allows u to use Router transformation to test data for multiple conditions. Router
transformation allows u route groups of data to transformation or target.
U can use XML data as a source or target.
Server Enhancements:
U can use the command line program pmcmd to specify a parameter file to run sessions or batches. This
allows you to change the values of session parameters, and mapping parameters and variables at runtime.
If you run the Informatica Server on a symmetric multi-processing system, you can use multiple CPUs to
process a session concurrently. You configure partitions in the session properties based on source qualifiers.
The Informatica Server reads, transforms, and writes partitions of data in parallel for a single session. This is
available for Power center only.
Informatica server creates two processes like load manager process, TM process to run the sessions.
Metadata Reporter: It is a web based application which is used to run reports against repository metadata.
U can copy the session across the folders and repositories using the copy session wizard in the informatica
server manager
with new email variables, you can configure post-session email to include information, such as the mapping
used during the session
108.What is incremental aggregation?
When using incremental aggregation, you apply captured changes in the source to aggregate calculations in
a session. If the source changes only incrementally and you can capture changes, you can configure the
session to process only those changes. This allows the Informatica Server to update your target
incrementally, rather than forcing it to process the entire source and recalculate the same calculations each
time you run the session.
109.What r the scheduling options to run a session?
U can schedule a session to run at a given time or interval, or u can manually run the session.
Different options of scheduling
Run only on demand: server runs the session only when user starts session explicitly
Run once: Informatica server runs the session only once at a specified date and time.
Run every: Informatica server runs the session at regular intervals as u configured.
Customized repeat: Informatica server runs the session at the data and times specified in the repeat dialog
box.

110.What is difference between stored procedure transformation and external procedure


transformation?
In case of stored procedure transformation procedure will be compiled and executed in a relational data
source. U need data base connection to import the stored procedure in to ur maping.Where as in external
procedure transformation procedure or function will be executed out side of data source. I.e. u need to make
it as a DLL to access in u r maping.No need to have data base connection in case of external procedure
transformation.
111.Explain about Recovering sessions?
If you stop a session or if an error causes a session to stop, refer to the session and error logs to determine
the cause of failure. Correct the errors, and then complete the session. The method you use to complete the
session depends on the properties of the mapping, session, and Informatica Server configuration.
Use one of the following methods to complete the session:
Run the session again if the Informatica Server has not issued a commit.
Truncate the target tables and run the session again if the session is not recoverable.
Consider performing recovery if the Informatica Server has issued at least one commit.
112.If a session fails after loading of 10,000 records in to the target. How can u load the records from
10001 th record when u run the session next time?
As explained above informatcia server has 3 methods to recovering the sessions. Use performing recovery
to load the records from where the session fails.
113.How to recover the standalone session?
A standalone session is a session that is not nested in a batch. If a standalone session fails, you can run
recovery using a menu command or pmcmd. These options are not available for batched sessions.
To recover sessions using the menu:
1. In the Server Manager, highlight the session you want to recover.
2. Select Server Requests-Stop from the menu.
3. With the failed session highlighted, select Server Requests-Start Session in Recovery Mode from the
menu.
To recover sessions using pmcmd:
1.From the command line, stop the session.
2. From the command line, start recovery.
114.If i done any modifications for my table in back end does it reflect in informatca warehouse or
maping desginer or source analyzer?
NO. Informatica is not at all concern with back end data base. It displays u all the information
that is to be stored in repository. If want to reflect back end changes to informatica screens,
again u have to import from back end to informatica by valid connection. And u have to replace the existing
files with imported files.
115.After draging the ports of three sources(sql server,oracle,informix) to a single source qualifier,
can u map these three ports directly to target?
NO.Unless and until u join those three ports in source qualifier u cannot map them directly.
116.What is Data cleansing..?
The process of finding and removing or correcting data that is incorrect, out-of-date, redundant, incomplete,
or formatted incorrectly.
117.How can we partition a session in Informatica?
The Informatica Power Center partitioning option optimizes parallel processing on multiprocessor
hardware
by
providing
a
thread-based
architecture
and
built-in
data
partitioning.
GUI-based tools reduce the development effort necessary to create data partitions and streamline ongoing
troubleshooting and performance tuning tasks, while ensuring data integrity throughout the execution
process. As the amount of data within an organization expands and real-time demand for information grows,
the Power Center Partitioning option enables hardware and applications to provide outstanding performance
and jointly scale to handle large volumes of data and users.

118.What is a time dimension? give an example.


Time dimension is one of important in Datawarehouse. Whenever u genetated the report , that time u
access all data from thro time dimension.
eg. Employee time dimension
Fields: Date key, full date, day of week, day , month, quarter, fiscal year.
119. Diff between informatica repository server & informatica server
Informatica Repository Server: Its manages connections to the repository from client application.
Informatica Server: Its extracts the source data, performs the data transformation, and loads the
transformed data into the target
120.Explain the informatica Architecture in detail ?
Informatica server connects source data and target data using native
odbc drivers again it connect to the repository for running sessions and retrieving metadata information
Source------>informatica server--------->target
|
|
REPOSITORY
121.Discuss the advantages & Disadvantages of star & snowflake schema?
In a star schema every dimension will have a primary key.
In a star schema, a dimension table will not have any parent table.
Whereas in a snow flake schema, a dimension table will have one or more parent tables.
Hierarchies for the dimensions are stored in the dimensional table itself in star schema.
Whereas hierarchies are broken into separate tables in snow flake schema. These hierarchies helps to drill
down the data from topmost hierarchies to the lowermost hierarchies.
star schema consists of single fact table surrounded by some dimensional table.In snowflake schema the
dimension tables are connected with some sub dimension table.
In star flake dimensional tables r denormalized,in snowflake dimension tables r normalized.
star schema is used for report generation ,snowflake schema is used for cube.
The advantage of snowflake schema is that the normalized tables r easier to maintain. it also saves the
storage space.
The disadvantage of snowflake schema is that it reduces the effectiveness of navigation across the tables
due to large no of joins between them.
122.What are main advantages and purpose of using Normalizer Transformation in Informatica?
Normalizer Transformation is used mainly with COBOL sources where most of the time data is stored in denormalized format. Also, Normalizer transformation can be used to create multiple rows from a single row of
data.
123.How to read rejected data or bad data from bad file and reload it to target?
Correction the rejected data and send to target relational tables using loadorder utility. Find out the rejected
data by using column indicatior and row indicator.
124.How do you transfert the data from data warehouse to flatfile?
You can write a mapping with the flat file as a target using a DUMMY_CONNECTION. A flat file target is built
by pulling a source into target space using Warehouse Designer tool.
125.At the max how many tranformations can be us in a mapping?
There is no such limitation to use this number of transformations. But in performance point of view using too
many transformations will reduce the session performance.
My idea is "if needed more transformations to use in a mapping its better to go for some stored procedure."

126.What is the difference between Normal load and Bulk load?


Normal Load: Normal load will write information to the database log file so that if any recorvery is needed it
is will be helpful. When the source file is a text file and loading data to a table,in such cases we should you
normal load only, else the session will be failed.
Bulk Mode: Bulk load will not write information to the database log file so that if any recorvery is needed we
can't do any thing in such cases. Comparatively Bulk load is pretty faster than normal load.
127.What is a junk dimension
A "junk" dimension is a collection of random transactional codes, flags and/or text attributes that are
unrelated to any particular dimension. The junk dimension is simply a structure that provides a convenient
place to store the junk attributes. A good example would be a trade fact in a company that brokers equity
trades.
128.Can we lookup a table from a source qualifer transformation-unconnected lookup?
No. we can't do.
I will explain you why.
1) Unless you assign the output of the source qualifier to another transformation or to target no way it will
include the field in the query.
2) Source qualifier doesnt have any variables fields to utilize as expression.
129.How to get the first 100 rows from the flat file into the target?
Task ----->(link) session (workflow manager)
Double click on link and type $$source success rows(parameter in session variables) = 100
it should automatically stops session.
130.Can we modify the data in flat file?
131.Difference between summary filter and details filter?
Summary Filter --- we can apply records group by that contain common values.
Detail Filter --- we can apply to each and every record in a database.
132.What are the difference between view and materialized view?
Materialized views are schema objects that can be used to summarize, precompiled, replicate, and distribute
data. E.g. to construct a data warehouse.
A materialized view provides indirect access to table data by storing the results of a query in a separate
schema object. Unlike an ordinary view, which does not take up any storage space or contain any data
133.Compare Data Warehousing Top-Down approach with Bottom-up approach?
Top down
ODS-->ETL-->Datawarehouse-->Datamart-->OLAP
Bottom up
ODS-->ETL-->Datamart-->Datawarehouse-->OLAP
Bottom approach is the best because in 3 tier architecture data tier is the bottom one.
134.Discuss which is better among incremental load, Normal Load and Bulk load?
Normal load is the best.
135.What is the difference between connected and unconnected stored procedures?
Unconnected:
The unconnected Stored Procedure transformation is not connected directly to the flow of the mapping. It
either runs before or after the session, or is called by an expression in another transformation in the
mapping.
Connected:
The flow of data through a mapping in connected mode also passes through the Stored Procedure
transformation. All data entering the transformation through the input ports affects the stored procedure. You
should use a connected Stored Procedure transformation when you need data from an input port sent as an
input parameter to the stored procedure, or the results of a stored procedure sent as an output parameter to
another transformation.

Run a stored procedure before or after your session.


Run a stored procedure once during your mapping, such as pre- or post-session.
Run a stored procedure every time a row passes through the Stored Procedure
transformation.
Run a stored procedure based on data that passes through the mapping, such as when
a specific port does not contain a null value.
Pass parameters to the stored procedure and receive a single output parameter.

Unconnected
Unconnected
Connected or
Unconnected

Unconnected
Connected or
Unconnected

Pass parameters to the stored procedure and receive multiple output parameters.
Note: To get multiple output parameters from an unconnected Stored Procedure
transformation, you must create variables for each output parameter. For details, see
Calling a Stored Procedure From an Expression.
Run nested stored procedures.
Call multiple times within a mapping.

Connected or
Unconnected

Unconnected
Unconnected

136.Differences between Informatica 6.2 and Informatica 7.0


In 7.0 introduce custom transformation and union transformation and also flat file lookup condition.
Features in 7.1 are:
1. Union and custom transformation
2. Lookup on flat file
3. Grid servers working on different operating systems can coexist on same server
4. We can use pmcmdrep
5. We can export independent and dependent rep objects
6. We ca move mapping in any web application
7. Version controlling
8. Data profiling.
137.What are the Differences between Informatica Power Center versions 6.2 and 7.1, also between
Versions 6.2 and 5.1?
Versioning
LDAP authentication
Support of 64 bit architectures.
138.What is the diff between Informatica powercenter server, repositoryserver and repository?
Power center server contains the scheduled runs at which time data should load from source to target
Repository contains all the definitions of the mappings done in designer.
139.How to create the staging area in your database?
A Staging area in a DW is used as a temporary space to hold all the records from the source system. So
more or less it should be exact replica of the source systems except for the load strategy where we use
truncate and reload options. So create using the same layout as in your source tables or using the Generate
SQL option in the Warehouse Designer tab.
140.What does the expression n filter transformations do in Informatica Slowly growing target
wizard?
EXPESSION transformation detects and flags the rows from source.Filter transformation filters the rows that
are not flagged and passes the flagged rows to the Update strategy transformation

141.In a filter expression we want to compare one date field with a db2 system field CURRENT
DATE.Our Syntax: datefield = CURRENT DATE (we didn't define it by ports, its a system field ), but
this is not valid (PMParser: Missing Operator)..
The db2 date format is "yyyymmdd" where as sysdate in oracle will give "dd-mm-yy" so conversion of db2
date format to local database date format is compulsory. Other wise u will get that type of error.
142.Briefly explian the Versioning Concept in Power Center 7.1?
In power center 7.1 use 9 Tem servers i.e. add in Look up. But in power center 6.x use only 8 tem servers.
And add 5 transformations. In 6.x any 17 transformations but 7.x uses 22 transformations.
143.How to join two tables without using the Joiner Transformation?
Its possible to join the two or more tables by using source qualifier. But provided the tables should have
relationship. When u drag n drop the tables u will getting the source qualifier for each table. Delete all the
source qualifiers. Add a common source qualifier for all. Right click on the source qualifier u will find EDIT
click on it. Click on the properties tab will find sql query in that u can write ur sql.
144.Can Informatica be used as a Cleansing Tool? If Yes, give example of transformations that can
implement a data cleansing routine.
Yes, we can use Informatica for cleansing data. Some time we use stages to cleansing the data. It depends
upon performance again else we can use expression to cleansing data.
For example an field X have some values and other with Null values and assigned to target feild where
target field is not null column, inside an expression we can assign space or some constant value to avoid
session failure.The input data is in one format and target is in another format, we can change the format in
expression.we can assign some default values to the target to represent complete set of data in the target.
145.How do you decide whether you need do aggregations at database level or at Informatica level?
It depends upon our requirement only. If you have good processing database you can create aggregation
table or view at database level else its better to use informatica. Here i'm explain why we need to use
informatica. what ever it may be informatica is a third party tool, so it will take more time to process
aggregation compared to the database, but in Informatica an option we called "Incremental aggregation"
which will help you to update the current values with current values +new values. No necessary to process
entire values again and again. Unless this can be done if nobody deleted that cache files. If that happened
total aggregation we need to execute on informatica also. In database we don't have Incremental
aggregation facility.
146.How do we estimate the depth of the session scheduling queue? Where do we set the number of
maximum concurrent sessions that Informatica can run at a given time?
147.How do we estimate the number of partitions that a mapping really requires? Is it dependent on
the machine configuration?
It depends upon the informatica version we r using. Suppose if we r using informatica 6 it supports only 32
partitions where as informatica 7 supports 64 partitions.
148.Suppose session is configured with commit interval of 10,000 rows and source has 50,000 rows.
Explain the commit points for Source based commit and Target based commit. Assume appropriate
value wherever required?
Source based commit will commit the data into target based on commit interval. So, for every 10,000 rows it
will commit into target.Target based commit will commit the data into target based on buffer size of the
target. i.e., it commits the data into target when ever the buffer fills. Let us assume that the buffer size is 6,
000.So, for every 6,000 rows it commits the data.
149.We are using Update Strategy Transformation in mapping how can we know whether insert or
update or reject or delete option has been selected during running of sessions in Informatica?
In Designer while creating Update Strategy Transformation uncheck "forward to next transformation". If any
rejected rows are there automatically it will be updated to the session log file.
Update or insert files are known by checking the target file or table only.
Operation
Constant
Numeric value
Insert
DD_INSERT
0
Update
DD_UPDATE
1
Delete
DD_DELETE
2
Reject
DD_REJECT
3

150.What is the procedure to write the query to list the highest salary of three employees?
SELECT sal
FROM (SELECT sal FROM my_table ORDER BY sal DESC)
WHERE ROWNUM < 4;
151.What is the limit to the number of sources and targets you can have in a mapping?
There is one formula..
no.of bloccks=0.9*( DTM buffer size/block size)*no.of partitions.
here no.of blocks=(source+targets)*2
the restriction is only on the database side. How many concurrent threads r u allowed to run on the db
server
152.Which is better among connected lookup and unconnected lookup transformations in
informatica or any other ETL tool?
Its not a easy question to say which is better out of connected, unconnected lookups. Its depends upon our
experience and upon the requirement.
When you compared both basically connected lookup will return more values and unconnected returns one
value. Conn lookup is in the same pipeline of source and it will accept dynamic caching. Unconn
lookup don't have that facility but in some special cases we can use Unconnected. if o/p of one lookup is
going as i/p of another lookup this unconnected lookups are favorable.
153.In Dimensional modeling fact table is normalized or denormalized?in case of star schema and
incase of snow flake schema?
In Dimensional modeling, Star Schema: A Single Fact table will be surrounded by a group of Dimensional
tables comprise of de- normalized data Snowflake Schema: A Single Fact table will be surrounded by a
group of Dimensional tables comprised of normalized dataThe Star Schema (sometimes referenced as star
join schema) is the simplest data warehouse schema, consisting of a single "fact table" with a compound
primary key, with one segment for each "dimension" and with additional columns of additive, numeric
facts.The Star Schema makes multi-dimensional database (MDDB) functionality possible using a traditional
relational database. Because relational databases are the most common data management system in
organizations today, implementing multi-dimensional views of data using a relational database is very
appealing. Even if you are using a specific MDDB solution, its sources likely are relational databases.
Another reason for using star schema is its ease of understanding. Fact tables in star schema are mostly in
third normal form (3NF), but dimensional tables in de-normalized second normal form (2NF). If you want to
normalize dimensional tables, they look like snowflakes (see snowflake schema) and the same problems of
relational databases arise - you need complex queries and business users cannot easily understand the
meaning of data. Although query performance may be improved by advanced DBMS technology and
hardware, highly normalized tables make reporting difficult and applications complex.The Snowflake
Schema is a more complex data warehouse model than a star schema, and is a type of star schema. It is
called a snowflake schema because the diagram of the schema resembles a snowflake.Snowflake schemas
normalize dimensions to eliminate redundancy. That is, the dimension data has been grouped into multiple
tables instead of one large table. For example, a product dimension table in a star schema might be
normalized into a products table, a Product-category table, and a product-manufacturer table in a snowflake
schema. While this saves space, it increases the number of dimension tables and requires more foreign key
joins. The result is more complex queries and reduced query performance.
Star schema--De-Normalized Dimensions
Snow Flake Schema-- Normalized Dimensions
154.What is difference between IIF and DECODE function?
You can use nested IIF statements to test multiple conditions. The following example tests for various
conditions and returns 0 if sales is zero or negative:
IIF( SALES > 0, IIF( SALES < 50, SALARY1, IIF( SALES < 100, SALARY2, IIF( SALES < 200, SALARY3,
BONUS))), 0 )
You can use DECODE instead of IIF in many cases. DECODE may improve readability. The following shows
how you can use DECODE instead of IIF :
SALES > 0 and SALES < 50, SALARY1,
SALES > 49 AND SALES < 100, SALARY2,
SALES > 99 AND SALES < 200, SALARY3,
SALES > 199, BONUS)

155.What are variable ports and list two situations when they can be used?
We have mainly tree ports Inport, Outport, Variable port. Inport represents data is flowing into
transformation. Outport is used when data is mapped to next transformation. Variable port is used when we
mathematical caluculations are required. If any addition i will be more than happy if you can share.
156.How does the server recognise the source and target databases?
By using ODBC connection.if it is relational.if is flat file FTP connection..see we can make sure with
connection in the properties of session both sources && targets.
157.How to retrive the records from a rejected file. explane with syntax or example?
There is one utility called "reject Loader" where we can findout the reject records.and able to refine and
reload the rejected records..
158.How to lookup the data on multiple tabels?
Using SQL override..we can lookup the Data on multiple tables.See in the properties.
159.What is the procedure to load the fact table.Give in detail?
We use the 2 wizards (i.e) the getting started wizard and slowly changing dimension wizard to load the fact
and dimension tables,by using these 2 wizards we can create different types of mappings according to the
business requirements and load into the star schemas(fact and dimension tables).
160.What is the use of incremental aggregation? Explain me in brief with an example?
Its a session option. When the informatica server performs incremental aggregator. it passes new source
data through the mapping and uses historical cache data to perform new aggregation calculations
incrementally. For performance we will use it.
161.How to delete duplicate rows in flat files source is any option in informatica?
Use a sorter transformation, in that u will have a "distinct" option make use of it.
162.How to use mapping parameters and what is their use?
Mapping parameters and variables make the use of mappings more flexible. and also it avoids creating of
multiple mappings. it helps in adding incremental data. Mapping parameters and variables has to create in
the mapping designer by choosing the menu option as Mapping ----> parameters and variables and the
enter the name for the variable or parameter but it has to be preceded by $$. and choose type as
parameter/variable, data type once defined the variable/parameter is in the any expression for example in
SQ transformation in the source filter properties tab. just enter filter condition and finally create a parameter
file to assign the value for the variable / parameter and configure the session properties. however the final
step is optional. if there parameter is npt present it uses the initial value which is assigned at the time of
creating the variable
163.In the concept of mapping parameters and variables, the variable value will be saved to the
repository after the completion of the session and the next time when u run the session, the server
takes the saved variable value in the repository and starts assigning the next value of the saved
value. for example i ran a session and in the end it stored a value of 50 to the repository.next time
when i run the session, it should start with the value of 70. not with the value of 51.
u can do one thing after running the mapping,, in workflow manager
Start-------->session.
Right click on the session u will get a menu, in that go for persistent values, there u will find the last value
stored in the repository regarding to mapping variable. Then remove it and put ur desired one, run the
session... i hope ur task will be done.
164.Significance of oracle 9i in informatica when compared to oracle 8 or 8i.
i mean how is oracle 9i advantageous when compared to oracle 8 or 8i when used in informatica
Its very easy Actually oracle 8i not allowed user defined data types but 9i allows and then blob,clob allow
only 9i not 8i and more over list partition is there in 9i only.
165.Can we use aggregator/active transformation after update strategy transformation?
We can use, but the update flag will not be remain. but we can use passive transformation

166.Why dimenstion tables are denormalized in nature ?


Because in Data warehousing historical data should be maintained, to maintain historical data means
suppose one employee details like where previously he worked, and now where he is working, all details
should be maintain in one table, if u maintain primary key it won't allow the duplicate records with same
employee id. so to maintain historical data we are all going for concept data warehousing by using surrogate
keys we can achieve the historical data(using oracle sequence for critical column).so all the dimensions are
marinating historical data, they are de normalized, because of duplicate entry means not exactly duplicate
record with same employee number another record is maintaining in the table
167.Why dimenstion tables are denormalized in nature ?
Because in Data warehousing historical data should be maintained, to maintain historical data means
suppose one employee details like where previously he worked, and now where he is working, all details
should be maintain in one table, if u maintain primary key it won't allow the duplicate records with same
employee id. so to maintain historical data we are all going for concept data warehousing by using surrogate
keys we can achieve the historical data(using oracle sequence for critical column).so all the dimensions are
marinating historical data, they are de normalized, because of duplicate entry means not exactly duplicate
record with same employee number another record is maintaining in the table.
168.In a sequential Batch how can we stop single session?
We have a task called wait event using that we can stop. We start using raise event.This is as per my
knowledge.
169.How do you handle decimal places while importing a flatfile into informatica?
While importing the flat file, the flat file wizard helps in configuring the properties of the file so that select the
numeric column and just enter the precision value and the scale. precision includes the scale for example if
the number is 98888.654, enter precision as 8 and scale as 3 and width as 10 for fixed width flat file.
170.If you are workflow is running slow in informatica. Where do you start trouble shooting and what
are the steps you follow?
When the work flow is running slowly u has to find out the bottlenecks in this order
Target
Source
Mapping
Session
System
171.If you have four lookup tables in the workflow. How do you troubleshoot to improve
performance?
There r many ways to improve the mapping which has multiple lookups.
1) We can create an index for the lookup table if we have permissions (staging area).
2) Divide the lookup mapping into two (a) dedicate one for insert means: source - target,, these r new rows .
Only the new rows will come to mapping and the process will be fast. (b) Dedicate the second one to
update: source=target, these r existing rows. Only the rows which exist already will come into the mapping.
3) We can increase the cache size of the lookup.
172.Can anyone explain error handling in informatica with examples so that it will be easy to explain
the same in the interview.
You can create some generalized transformations to handle the errors and use them in your mapping. For
example for data types create one generalized transformation and include in your mapping then you will
know the errors where they are occurring.
173.How do I import VSAM files from source to target. Do I need a special plug-in?
174.Differences between Normalize and Normalizer transformation.?
Normalizer: It is a transformation mainly using for Cobol sources,Its change the rows into columns and
columns into rows
Normalization: To remove the redundancy and inconsistency.

175.What is IQD file?


IQD file is nothing but Impromptu Query Definition, This file is maily used in Cognos Impromptu tool after
creating a imr( report) we save the imr as IQD file which is used while creating a cube in power play
transformer. In data source type we select Impromptu Query Definition.
176.What is data merging, data cleansing, sampling?
Cleansing: ---TO identify and remove the retundacy and inconsistency
Sampling: just sample the data through send the data from source to target
177.How to import oracle sequence into Informatica?
CREATE ONE PROCEDURE AND DECLARE THE SEQUENCE INSIDE THE PROCEDURE,FINALLY
CALL THE PROCEDURE IN INFORMATICA WITH THE HELP OF STORED PROCEDURE
TRANSFORMATION.
178. What are the steps required for type2 dimension/version data mapping. how can we implement
it?
Go to mapping designer in it go for mapping select wizard in it go for slowly changing dimension Here u will
find a new window their u need to give the mapping name source table target table and type of scd then if
select finish scd 2 mapping is created go to ware designer and generate the table then validate the mapping
in mapping designer save it to repository run the session in workflow manager later update the source table
and re run again u will find the difference in target table.
179.With out using Update strategy and sessions options, how we can do the update our target
table?
Using "update override" in target option.
180.Two relational tables are connected to SQ Trans,what are the possible errors it will be thrown?
We can connect two relational tables in one sq Transformation. No errors will be performing.
181.What is the best way to show metadata(number of rows at source, target and each
transformation level, error related data) in a report format?
When your workflow get completed go to workflow monitor right click the session .then go to transformation
statistics there we can see number of rows in source and target.if we go for session properties we can see
errors related to data.
182.If u had to split the source level key going into two seperate tables. One as surrogate and other
as primary. Since informatica does not gurantee keys are loaded properly(order!) into those tables.
What are the different ways you could handle this type of situation?
183.How to append the records in flat file(Informatica) ? Where as in Datastage we have the options
i) overwrite the existing file
ii) Append existing file
This is not there in Informatica v 7. but heard that its included in the latest version 8.0 where u can append
to a flat file. Its about to be shipping in the market.
184.What are partition points?
Partition points mark the thread boundaries in a source pipeline and divide the pipeline into stages.
185.What are cost based and rule based approaches and the difference?
Cost based and rules based approaches are the optimization techniques which are used in related to
databases, where we need to optimize a sql query. Basically Oracle provides Two types of Optimizers
(indeed 3 but we use only these two techniques. bcz the third has some disadvantages.)When ever you
process any sql query in Oracle, what oracle engine internally does is, it reads the query and decides which
will the best possible way for executing the query. So in this process, Oracle follows these optimization
techniques.
1. cost based Optimizer (CBO): If a sql query can be executed in 2 different ways (like may have path 1 and
path2 for same query), then What CBO does is, it basically calculates the cost of each path and the
analyses for which path the cost of execution is less and then executes that path so that it can optimize the
query execution.
2. Rule base optimizer (RBO): this basically follows the rules which are needed for executing a query. So
depending on the number of rules which are to be applied, the optimizer runs the query.

Use:
If the table you are trying to query is already analyzed, then oracle will go with CBO. If the table is not
analyzed, the Oracle follows RBO.
For the first time, if table is not analyzed, Oracle will go with full table scan.
186.What is mystery dimension?
Using Mystery Dimension ur maintaining the mystery data in ur Project.
187.What is difference b/w Informatica 7.1 and Abinitio?
In Informatica there is the concept of co-operating system, which makes the mapping in parallel fashion
which is not in Informatica.
188.Can i start and stop single session in concurrent batch?
Just right click on the particular session and going to recovery option.
189.What is difference between lookup cache and uncached lookup?
190. Can i run the mapping with out starting the informatica server?
The difference between cache and uncached lookup is when you configure the lookup transformation cache
lookup it stores all the lookup table data in the cache when the first input record enter into the lookup
transformation, in cache lookup the select statement executes only once and compares the values of the
input record with the values in the cache but in uncache lookup the the select statement executes for each
input record entering into the lookup transformation and it has to connect to database each time entering the
new record.
191.What is the difference between stop and abort?
Stop: If the session u want to stop is a part of batch you must stop the batch,if the batch is part of nested
batch, Stop the outer most batch
Abort: You can issue the abort command; it is similar to stop command except it has 60 second time out.
If the server cannot finish processing and committing data with in 60 sec
192.Can we run a group of sessions without using workflow manager?
Its Possible using pmcmd Command with out using the workflow Manager run the group of session.
193.How to perform a "Loop Scope / Loop condition" in an Informatica
program ? Give me few examples.
194.If a session fails after loading of 10,000 records in to the target.How can u load the records from
10001 th record when u run the session next time in informatica 6.1?
Using performance recovery option.
195.I have an requirement where in the columns names in a table (Table A) should appear in rows of
target table (Table B) i.e. converting columns to rows. Is it possible through Informatica? If so, how?
If data in tables as follows
Table A
Key-1 char(3);
table A values
_______
1
2
3
Table B
bkey-a char(3);
bcode char(1);
table b values
1T
1A
1G
2A

2T
2L
3A
and output required is as
1, T, A
2, A, T, L
3, A
the SQL query in source qualifier should be
select key_1,
max(decode( bcode, 'T', bcode, null )) t_code,
max(decode( bcode, 'A', bcode, null )) a_code,
max(decode( bcode, 'L', bcode, null )) l_code
from a, b
where a.key_1 = b.bkey_a
group by key_1
/
196.What is meant by complex mapping?
Complex mapping means involved in more logic and more business rules. Actually in my project complex
mapping is In my bank project, I involved in construct a 1 data ware house Many customer is there in my
bank project, They r after taking loans relocated in to another place that time I feel to difficult maintain both
previous and current addresses in the sense i am using scd2.This is an simple example of complex
mapping.
197.Explain use of update strategy transformation?
Maintain the history data and maintain the most recent changes data.
198.What are mapping parameters and varibles in which situation we can use it?
Mapping parameters have a constant value through out the session whereas in mapping variable the values
change and the informatica server saves the values in the repository and uses next time when u run the
session.
199.What is work let and what use of work let and in which situation we can use it?
Worklet is a set of tasks. If a certain set of task has to be reused in many workflows then we use worklets.
To execute a Worklet, it has to be placed inside a workflow. The use of worklet in a workflow is similar to the
use of mapplet in a mapping.
200.what is difference between dimention table and fact table and what are different dimention tables
and fact tables?
In the fact table contain measurable data and less columns and many rows, its contain primary key
Different types of fact tables:
Additive, non additive, semi additive
In the dimensions table contain textual description of data and also contain many columns, less rows
Its contain primary key.
201.How do you configure mapping in informatica?
You should configure the mapping with the least number of transformations and expressions to do the most
amount of work possible. You should minimize the amount of data moved by deleting unnecessary links
between transformations.
For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup transformations),
limit connected input/output or output ports. Limiting the number of connected input/output or output ports
reduces the amount of data the transformations store in the data cache.
You can also perform the following tasks to optimize the mapping:
Configure single-pass reading.
Optimize datatype conversions.
Eliminate transformation errors.
Optimize transformations.
Optimize expressions. You should configure the mapping with the least number of transformations
and expressions to do the most amount of work possible. You should minimize the amount of data
moved by deleting unnecessary links between transformations.

For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup
transformations), limit connected input/output or output ports. Limiting the number of connected
input/output or output ports reduces the amount of data the transformations store in the data cache.
You can also perform the following tasks to optimize the mapping:
o Configure single-pass reading.
o Optimize datatype conversions.
o Eliminate transformation errors.
o Optimize transformations.
o Optimize expressions.
202.Can i use a session Bulk loading option that time can i make a recovery to the session?
No, why because in bulk load u wont create redo log file, when u normal load we create redo log file, but in
bulk load session performance increase.
203.What is lookup transformation and update strategy transformation and explain with an example?
Look up transformation is used to lookup the data in a relation table,view,Synonym and Flat file.The
informatica server queries the lookup table based on the lookup ports used in the transformation.It compares
the lookup transformation port values to lookup table column values based on the lookup conditionBy using
lookup we can get the realated value,Perform a caluclation and Update SCD.
Two types of lookups
Connected
Unconnected
Update strategy transformation This is used to control how the rows are flagged for insert, update ,delete or
reject. To define a flagging of rows in a session it can be insert, Delete, Update or Data driven.
In Update we have three options
Update as Update
Update as insert
Update else insert.
204.What is the difference between Power Centre and Power Mart? What is the procedure for
creating Independent Data Marts from Informatica 7.1?
Power center
Power mart
No. of repository
n No.
n No.
Applicability
high end WH
low-minded range WH
Global repository
supported
not supported
Local repository
supported
supported
ERP support
available
not available
205.In the source, if we also have duplicate records and we have 2 targets, T1- for unique values and
T2- only for duplicate values. How do we pass the unique values to T1 and duplicate values to T2
from the source to these 2 different targets in a single mapping?
This is not a right approach friends. There is a good practice of identifying duplicates. Normally when you
ask someone how to identify a duplicate record in informatica, they say "Use aggregator transf". Well you
can just get a count from this, but not really identify which record is a duplicate. If it is RDBMS, you can
simply write a query "select ... from ...group by <key fields> having count (*) > 1. Great! But what if the
source is a flat file? you can use an aggregate and get the count of it. Then you will filter and wanted to
make sure it reached the T1 and T2 tgt's appropriately.This would be the easiest way.Use a sorter
transformation. Sort on key fields by which u want to find the duplicates. then use an expression
transformation.
Example:
Example:
field1-->
field2-->
SORTER:
field1 --ascending/descending
field2 --ascending/descending
Expression:
--> field1
--> field2
<--> v_field1_curr = field1
<--> v_field2_curr = field2

v_dup_flag = IIF(v_field1_curr = v_field1_prev, true, false)


o_dup_flag = IIF(v_dup_flag = true, 'Duplicate', 'Not Duplicate'
<--> v_field1_prev = v_field1_curr
<--> v_field2_prev = v_field2_curr
use a Router transformation and put o_dup_flag = 'Duplicate' in T2 and 'Not Duplicate' in T1.
Informatica evaluates row by row. So as we sort, all the rows come in order and it will evaluate based on the
previous and current rows.
206.How do you create a mapping using multiple lookup transformation?
207Can any one explain real time complain mappings or complex transformations in
Informatica.Specially in Sales Domain?
Most complex logic we use is demoralization. We dont have any Demoralizer transformation in INformatica.
So we will have to use an aggregator followed by an expression. Apart from this, we use most of the
complex city in expression transformation involving lot of nested IIF's and Decode statements...another one
is the union transformation and joiner.
208. About Informatica Power center 7:
1) I want to Know which mapping properties can be overridden on a Session Task level.
2) Know what types of permissions are needed to run and schedule Work flows.
You can override any properties other than the source and targets. Make sure the source and targets exist in
ur db if it is a relational db. If it is a flat file, you can override its properties. You can override sql if its a
relational db, session log, DTM buffer size, cache sizes etc.
209.Can anyone explain about incremental aggregation with an example?
When you use aggregator transformation to aggregate it creates index and data caches to store the data
1.Of group by columns 2. Of aggregate columns the incremental aggregation is used when we have
historical data in place which will be used in aggregation incremental aggregation uses the cache which
contains the historical data and for each group by column value already present in cache it add the data
value to its corresponding data cache value and outputs the row, in case of a incoming value having no
match in index cache the new values for group by and output ports are inserted into the cache.
210.Informatica Live Interview Questions?
Explain grouped cross tab?
Explain reference cursor
What are parallel query's and query hints
What is meta data and system catalog
What is factless fact schema
What is confirmed dimension
Which kind of index is preferred in DWH
Why do we use DSS database for OLAP tools
211.Can any body write a session parameter file which will change the source and targets for every
session. i.e. different source and targets for each session run.
You are supposed to define a parameter file. And then in the Parameter file, you can define two parameters,
one for source and one for target.
Give like this for example:
$Src_file = c:\program files\informatica\server\bin\abc_source.txt
$tgt_file = c:\targets\abc_targets.txt
Then go and define the parameter file:
[folder_name.WF:workflow_name.ST:s_session_name]
$Src_file =c:\program files\informatica\server\bin\abc_source.txt
$tgt_file = c:\targets\abc_targets.txt
If its a relational db, you can even give an overridden sql at the session level...as a parameter. Make sure
the sql is in a single line.
212.How do we remove the staging area?
213.What is polling?
It displays update information about the session in the monitor window.

214.What is Transaction?
A transaction can be defined as DML operation means it can be insertion, modification or deletion of data
performed by users/ analysts/applicators.
215.What are the various test procedures used to check whether the data is loaded in the backend,
performance of the mapping, and quality of the data loaded in INFORMATICA?
216.1)What are the various test procedures used to check whether the data is loaded in the backend,
performance of the mapping, and quality of the data loaded in INFORMATICA.
2) What are the common problems developers face while ETL development?
217.What happens if you try to create a shortcut to a non-shared folder?
It only creates a copy of it..
218.In a joiner transformation, you should specify the source with fewer rows as the master source.
Why?
In joinner transformation informatica server reads all the records from master source builds index and data
caches based on master table rows. after building the caches the joiner transformation reads records from
the detail source and perform joins.
219.If you want to create indexes after the load process which transformation you choose?
a) Filter Tranformation
b) Aggregator Tranformation
c) Stored procedure Tranformation
d) Expression Tranformation
Stored procedure transformation.
220.Where is the cache stored in informatica?
Cache stored in informatica is in informatica server.
221.How to get two targets T1 containing distinct values and T2 containing duplicate values from
one source S1?
222.What will happen if you are using Update Strategy Transformation and your session is
configured for "insert"?
What are the types of External Loader available with Informatica?
If you have rank index for top 10. However if you pass only 5 records, what will be the output of such
a Rank Transformation?
223.What are the real time problems generally come up while doing/running mapping/any
transformation?can any body explain with example.
224.Can batches be copied/stopped from server manager?
225.What is rank transformation? where can we use this transformation?
Rank transformation is used to find the status. Ex if we have one sales table and in this if we find more
employees selling the same product and we are in need to find the first 5 0r 10 employee who is selling
more products. we can go for rank transformation.
226.What is exact use of 'Online' and 'Offline' server connect Options while defining Work flow in
Work flow monitor? . The system hangs when 'Online' server connect option. The Informatica is
installed on a Personal laptop.
227.How can you delete duplicate rows with out using Dynamic Lookup? Tell me any other ways
using lookup delete the duplicate rows?

Business Objects
BO Designer
1.What is Cardinality?
Expresses the minimum and the maximum number of instances of an entity B that can be associated with
an instance of an entity A. The minimum and the maximum number of instances can be equal to 0, 1, or N.
2.What is Cartesian product?
A situation in which a query includes two or more tables that are not linked by a join. If executed, this type of
query retrieves all possible combinations between each table and may lead to inaccurate results.
3.What is Class?
A class is a logical grouping of objects within a universe. In general, the name of a class reflects a business
concept that conveys the category or type of objects. For example, in a universe pertaining to human
resources, one class might be Employees.A class can be further divided into subclasses. In the human
resources universe, a subclass of the Employees class could be Personal Information.As designer, you are
free to define hierarchies of classes and subclasses in a model that best reflects the business concepts of
your organization.
4. What is Condition?
A component that controls the type and the amount of data returned by a specific object in a query. A
condition created in the Designer module is referred to as a predefined condition.
5.What is Connection?
Set of parameter that provides access to an RDBMS. These parameters include system information such as
the data account, user identification, and path to the database. Designer provides three types of
connections: secured, shared, and personal.
6.What is Context?
A method by which Designer can decide which path to choose when more than one path is possible from
one table to another in the universe.
7.What is Detail object?
An object qualified as a detail provides descriptive data about a dimension object. A detail object cannot be
used in drill down analysis.
8.What is Dimension object?
An object being tracked in multidimensional analysis; the subject of the analysis. Dimensions are organized
into hierarchies.
9.What is Document domain?
The area of the repository that stores documents, templates, scripts, and lists of values.
10.What is Drill?
There can be 3 types of Drill Analysis- Drill Down,Drill Up and Drill Through.Within the same universe one
can Drill Up/Down like Country-State-City;of course with facts that relate to the same grain.A Drill through is
possible when we can link different Data Marts like Profitability as defined by details of Asset, Liability,
Income and Expense.
11.What is Equi-join?
A join based on the equality between the values in the column of one table and the values in the column of
another. Because the same column is present in both tables, the join synchronizes the two tables.
12.What is Enterprise mode?
A work mode whereby a designer creates universes in an environment with a repository. The mode in which
a universe is saved determines whether other designers are able to access it. By default, a universe is
saved in the mode in which the designer is already working.

13.What is Hierarchy?
An ordered series of related dimensions used for multidimensional analysis.
14.What is Join?
A relational operation that causes two tables with a common column to be combined into a single table.
Designer supports equi-joins, theta joins, outer joins, and shortcut joins.
15.What is List of values?
A list of values contains the data values associated with an object. These data values can originate from a
corporate database, or a flat file such as a text file or Excel file. In Designer you create a list of values by
running a query from the Query Panel. You can then view, edit, purge, refresh and even export this file. A list
of values is stored as an .lov file in a subfolder of the UserDocs folder.
16.What is Loop?
A situation that occurs when more than one path exists from one table to another in the universe.
17.What is Measure object?
An object that is derived from an aggregate function. It conveys numeric information by which a dimension
object can be measured.
18.What is Object?
A component that maps to data or a derivation of data in the database. For the purposes of multidimensional
analysis, an object can be qualified as a dimension, detail, or measure. Objects are grouped into classes.
19.What is Offline mode?
The work mode in which the designer works with universes stored locally.
20.What is Online mode?
The work mode appropriate for a networked environment in which the general supervisor has set up a
repository.
21.What is Outer join?
A join that links two tables, one of which has rows that do not match those in the common column of the
other table.
22.What is Personal connection?
A personal connection is used to access resources such as universes or documents. It can be used only by
the user who created it. Information about a personal connection is stored in both the PDAC.LSI and
PDAC.SSI files; its definition is static and cannot be modified.
23.What is Qualification?
A property of an object that determines how it can be used in multidimensional analysis. An object can be
qualified as one of three types: a dimension, detail or measure.
24.What is Query?
In Designer a query is a technique for creating or modifying a list of values associated with an object. From
the Query Panel, a designer builds a query from the classes, objects, and conditions of a universe. In the
BusinessObjects User module, a query is a type of data provider. An end user builds a query from a
universe, and then runs the query to generate a BusinessObjects report.
25.What is Quick Design?
A wizard in the Designer module that provides guided instructions for creating a basic universe. It lets a
designer name a universe, set up a connection to a database, select strategies, create classes and objects,
as well as generate joins with cardinalities.
26.What is Repository?
A centralized set of relational data structures stored in a database. It enables BusinessObjects users to
share resources in a controlled and secured environment. The repository is made up of three domains: the
security domain, the universe domain, and the document domain.

27.What is Secured connection?


A secured connection is used to access universes or documents that may be restricted or confidential. It can
be shared by several authorized users. Stored in the repository, the definition of a secured connection is
updated dynamically.
28.What is Shared connection?
A shared connection is used to access common resources such as universes or documents. It can be used
by several users. Information about a shared connection is stored in a SDAC.LSI or SDAC.SSI file; its
definition is updated dynamically.
29.What is Shortcut join?
A join that links two tables by bypassing one or more other tables in the universe.
30.What is Strategy?
Scripts that automatically extract structural information about tables, columns, joins, or cardinalities from a
database. Designer provides default strategies but a designer can also create strategies. These are referred
to as external strategies.
31.What is Structure pane?
The graphical component within the main Designer window that shows the schema of the universe. It
reflects the underlying database structure of the universe.
32.What is Subclass?
A component within a class that groups objects. A subclass can itself contain other subclasses or objects.
33.What is Table Browser?
The graphical component within the main Designer window that lets you create the classes and objects of
the universe from the tables and columns of a database.
34.What is Theta join?
A join that links tables based on a relationship other than equality between two columns.
35.What is Universe?
A universe is a business-oriented mapping of the data structure found in databases: tables, columns, joins,
etc. It can represent any specific application, system, or group of users. For example, a universe can relate
to a department in a company such as marketing or accounting.
36.What is Universe domain?
The area of the repository that holds exported universes. The universe domain makes it possible to store,
distribute, and administrate universes. There may be multiple universe domains in a repository.
37.What is Universe pane?
The graphical component within the main Designer window that displays a hierarchical view of the classes,
objects, and conditions in a universe. Two radio buttons on the lower edge of the pane filter the display of
the components. One button, the Classes/Objects filter, provides a view of the classes and objects. The
other button, Classes/Conditions, provides a view of the classes and conditions.
38.What is User object?
An object created by the end user from the Business Objects User module. A user object can be inserted
into a universe by the designer.
Insert a user object
1 Select the User Objects command from the Insert menu.
The Insert User Objects dialog box is displayed.
2 Click a user object file.
A user object file has a .udo extension
3 Click the Open button.
39.What is Workgroup mode?
A work mode whereby a designer creates universes in an environment without a repository. The mode in
which a universe is saved determines whether other designers are able to access it. By default, a universe is
saved in the mode in which the designer is already working.

40.What is Designer?
Designer is a Business Objects IS module used by universe designers to create and maintain universes.
Universes are the semantic layer that isolates end users from the technical issues of the database structure.
Universe designers can distribute universes to end users by moving them as files through the file system, or
by exporting them to the repository.
41.How do you design a universe?
The design method consists of two major phases. During the first phase, you create the underlying database
structure of your universe. This structure includes the tables and columns of a database and the joins by
which they are linked. You may need to resolve loops which occur in the joins using aliases or contexts. You
can conclude this phase by testing the integrity of the overall structure.During the second phase, you can
proceed to enhance the components of your universe. You can also prepare certain objects for
multidimensional analysis. As with the first phase, you should test the integrity of your universe structure.
You may also wish to perform tests on the universes you create from the BusinessObjects User module.
Finally, you can distribute your universes to users by exporting them to the repository or via your file system.
For a universe based on a simple relational schema, Designer provides Quick Design, a wizard for creating
a basic yet complete universe. You can use the resulting universe immediately, or you can modify the
objects and create complex new ones. In this way, you can gradually refine the quality and structure of your
universe.
42.What are the precautionary measures you will take in the project?
43.What is the drill up , drill down, drill by , drill trough ?
Drill up: UP-one level
Drill down: DOWN-one level
Drill by: selection of level Hierarchy
Drill trough: Hierarchy to another Hierarchy
44.Explain the SQL Queries activated to data base from Data provider in BO ?
BO automatically generates the SQL query when objects are selected into query panel. When you run the
query, the query is processed to database based on your connectivity. For ex: If you run query using, Local
Machine Full Client (BO reporter), the query directly connects to database through middleware
Full Client <---> Database
If you run query using Web, the web browser will connects to Web Server and Web server will process
request to Database.
WEBI <--->Web Server <---> Database
45.What are steps to be taken care to create a good Universe?
1)make the joins with optimization
2)reduce the creation user objects more in the universe
3)class should not be more than 60
4)try to use aggregate ware on measure objects
46.What are the steps to taken care to improve the Report performance?
In DESIGNER Level
1)eliminate the unnecessary joins
2)use conditions as much as at the database level
3)edit the SQL query in the Query Panel as per requirement
In REPORTER level
1)eliminate the filters as much as possible
2)try to reduce the user variables .
47.How can we achieve Correlated sub-query in Designer?can anyone help me in this regard.
Right click on any object, go to the properties. Specify the query in select and put the next query in where
clause, like select COLNAME from TABNAME1 where COLNAME IN (select colname2 from tab2)

48.What is broadcast agent?


Broadcast agent is used for scheduling or bursting the reports. thru this, you can schedule your reports so
that end-users can access reports thru internet. you have broadcast administration console here thru which,
you can schedule reports and watch the tasks even. It is an administrator tool of Business Object.
49.How to connect one to one report?
You can do this in different methods:
In ASP environment..
Universe level: create an object with Object format: HTML option enabled,
SELECT :< A href="Server:PortNo/wiasp/scripts/opendocument.asp?....etc..." >
Report Level: use Hyperlink function
Hyperlink("Server:PortNo/wiasp/scripts......etc..........." , "TaxtAsLink")
For JSP environment, replace ASP with JSP.
Any of the above methods will create a Hyperlink in the report pointing to another report. The
arguments/prompts for this new report can be passes from the original report.
50.What are the diffrences between Business Objects 5.1 & 6.5 ?
Main difference between BO 5.1 and 6.5 is Web Intelligence (info view).Using BO 5.1 info view. We can only
view the reports that are created using USER module. We can't modify those reports. Using BO 6.5 info
view. We can view and create the reports. We need not have BO installed on out systems to do this.
51.How do you connect the facts and dimensions in the tables?
1. Smart Matching columns
2. Manually you can link
52.I want to restrict top 26 rows ...in report is there any rownum object from which i can do this?
Restrict Top 26 Rows: Depends on how you want to do it.
In a report: Drag the Dimension and measures into report. Click on the Dimension object and use the RANK
button on the tool bar. You can specify the Top 26 or Bottom (Depending on the requirement) values of a
measure. It also provides some calculation like Sum/Percentage.If you want to restrict the Top 26 in the
query itself, you might have to create an object in the universe. May have to use the analytical functions
based on which you need to Rank the Data. For ex: use the Designer function : RANK() OVER (PARTITION
BY () ORDER BY ())
53.What is the local filter and global filter in bo's?
Local Filter >>Specific to a block (Table/cross tab/ chart etc.)
Global Filter >> Specific to a TAB in a report
If a report has multiple Tabs, each can have its own Local and Global filters.
54.What is pragma?
A PRAGMA statement appears in a PL/SQL block. It is not an executable statement, but it provides
instructions to the compiler.
55.Could any one explain me how to create cascading prompts?
If you have source system > source location field. double click source system > edit properties menu
will display, click properties tab and click edit you will get query pane there you pull down source
location object in conditon pane and click inlist and click type new prompt type there "XXXXX" and save that.
Now you click display button first source system lov will display and you select any one, based on that
selection the source location will display. This is cascading prompt.
56.Which command using Query Analyzer will give you the version of SQL server and operating
system?
57.Product date sales
p1 20/Jan/1999 100
p1 20/Feb/1999 200
p1 20/apl/1999 100
p1 20/DEC/1999 50
p1 20/Jan/2000 100

p1 20/mar/2000 500
now i want query like product ,sum of actual year ,sum of business year
p1 450 750 here actual year means 1 Jan 1999 to 31 st Jan 1999
business year means 1 st APR 1999 to 31 st mar 200
58.What is slicing and dicing in business objects?
Slice & Dice is facility in BO. We can enables change the positions of data in Report. Here in Bo we slice &
dice panel by using this we can create cross tables and master details tables.
59.How to link two different universes ?
For link 2 universes we hae 2 approaches..
1. Go through Edit-->links
2. Go through parameters of Universes there is one tab like Links, by using those we can link the 2
universes...
60.What's the Functional & Architectural Differences between Business Objects & Web Intelligence
Reports?
61.How to Implement the the Built-in Strategy Script in BO Designer?
62.What's is the Guidelines to Build Universe with Better Performance? R Performance tuning Issues
of Universes?

Business Objects
1.What is a bo repository?
Generally Repository is the Metadata
BO 5.0 Repository create/maintain 50 tables, they are distributed as follows
25 tables for Security domain
24 Table for Universe domain
1 Table for Document Domain
2.Give the notes and functionalities of cascading prompts,@script in business objects?
Syntax
@Prompt ('message', ['type'], [lov], [MONO|MULTI], [FREE|CONSTRAINED])
where message is the text of a message within single quotes. type can be one of the following: 'A' for
alphanumeric, 'N' for number, or 'D' for date. can be either a list of values enclosed in brackets (each value
must be within single quotes and separated by commas) or the name of a class and object separated by a
backslash and within single quotes. MONO means that the prompt accepts only one value. MULTI means
that the prompt can accept several values. FREE refers to free input as opposed to CONSTRAINED, which
means that the end user must choose a value suggested by the prompt.
Description
Is used to create an interactive object. In the Query Panel, this type of object causes a message to appear.
This
message
prompts
the
end
user
to
enter
a
specific
value.
Note
The last four arguments are optional; however, if you omit an argument you must still enter the commas as
separators.
Example
In Where Clause:
City.city IN @Prompt ('Choose City', 'A', {'Chicago', 'Boston', 'New York'}, MULTI, FREE)
In the Query Panel, the object prompts the end user to choose a city.
3.When to use local filter and when to use global Filter?
Local Filter is to single report to which it is create, but a global filter is to all the reports which consists of that
column
4.What are the user requirements in a universe?
Database connections, key column, joins and check for loop if you need measures, metrics.
5. I have three predefined prompts. In a report it will come randomly. How they will come in a
specified format?
The Prompts will appear in the alphabetical order.To make them appear in the order of our requirement,
need to prefix a numerical with the prompt.
5.Whats universal join in BOs?
The level of join between two universes with a matching column.
6.Can we apply Rank and Sort at a time on a single report?
No we can't apply rank and sort at a time on one object in one single report.If we try to apply , BO asks if
you want to over write the previous condition.
7.What is difference between custom hierarchy and report based hierarchy?
By default one class having one hierarchy ie called report hierarchy. Custom hierarchy we can create in
designer according our req.
8.What is the multi value error ?Is there any types of Error in BO?
You will get the Multi Value Error when you are trying to retrivr mutiple values into the cell.
Ex: When u r tying to Insert cell into report and trying to assign a column to it which will have
multiple values in it. (In A single cell you cant show multiple values)
9.How many ways we test the universe & Report?
By doing integrity check we can test universe & By coping report query and run in backend(oracle,sql
server...) we can test the data by comparing both.

10.How achieve the many to many relation ship in Business Objects


If you have two tables with many to much relation ship, take the keys of those tables and form a 1-1
relationship table. Then you can connect both the existing tables to the newly formed table using the key.
11.How to filter repetitive values in the report to get distinct values.
Example: In any database we can use distinct command But My Report source is flatfile.Therfore,I
cann\'t manipulate source system.
Is there any command or function to filter repeating records/values in the report.
Using the eliminate duplicates option is one way.
In a clightly different scenario, duplicates could come up due to the design structure of the report. e.g.
In some specific cases, one may want to display a value in one field (say in a folded report) but multiple
values come up (which could give you a COMPUTATION ERROR). To avoid such situation, one can define
a variable which would e the max of the original field required. That ways the same value is fetched and also
only one value is returned.
12.What is the difference between tabular report& cross tab report?
Tabular reports are a 2D format and Cross tab reports is a 3D format. Most of the operational/formatted
reports will be created in tabular form and most of the analytical reports are created in the cross tab report
form (Product VS Year VS Sales amount)
13.Can you add the Cascading prompts in Business Objects? (Based on Country prompt selection
you need to provide the option to select Sate/Province in the next State prompt. That means
dynamically refresh the data load in the state list values)
14. My query takes 30 min to refresh ...there is any performance tunning is there to reduce refresh
time?
Make sure to have the right context - apply right indexes on the tables.
15.What is the difference between condition & filter?
Conitions: these retrievals the data based on the condition (universe level)
Filter: it brings the data and filter; performance will be slow (report level)
16.How can we do load testing for Webi Reports? (Example if I have have to test one web report for
Concurrent users refreshing report at one time)
17.What is the diferrence between slice and dice & drill through?
Slice and dice is you are going to analyze the data in different angle, but drill through is like using a single
value you are going to grain the data.
18.What is the difference between slice and dice and cross tab report?
Cross tab report is like at the intersection of 2 dimensions measure is displayed.like if we have two
dimensions year and product and a measure revenu,revenue is displayed at the intersection of year and
product.
19.How we drill up week to Month?
In data Warehouse when u changes + symbol to - , u would c all the aggregation level in that + this is called
rolling down. So if u again click over that - it would again compressed back to +. This is call drill up.
Your Aggregation levels would b like
Year>Month>Week>Day>Hour>Minutes>Seconds.
So by changing the symbol from - to + at month aggregation level u can drill up.
20.How to break a Date field into Year, month, week, day?(For example: I have a field INVOICE_DATE.
Now i want to break it as Year, month..... I.e. in time dimension.) DO i need to have some changes in
Universe?
Using date functions separate the date, day, week, year, quarter, decade, holiday and more, and then you
can get your answers.
21.What is the dense rank?

Potrebbero piacerti anche