Informatica Ques Part 2
Informatica Ques Part 2
Informatica Ques Part 2
Is it passive or active when check and uncheck the box of DISTINCT in Sorter
transformation? why?
Answer
#1
Sorter Transformation is passive when the DISTINCT option is unchecked bcoz it
wont change the number of records passes through it. Once the DISTINCT option is
checked it
may change the no of records if the records has duplicates.
Answer
#2
It is Active transformation.
If you configure the Sorter transformation for distinct output rows, the Mapping
Designer configures all ports as part of the sort key. When the PowerCenter Server
runs the session, it discards duplicate rows compared during the sort operation.
Answer
#4
I sorter is an Active Transformation,I u check the Distinct option it will output
only the distint rows from the source .
Answer
#5
if we check the box distinct means u are eliminating the duplicate record.so here
if we check the distinct while sorting its eliminating the duplicate records and
after its
sorts so the number of records in target are less compared to source for that it
is calles as active. if we uncheck
this all the records are transforms into target table so it
is passive.
Question
if i have records like these
(source table)
rowid name
10001 gdgfj
10002 dkdfh
10003 fjfgdhgjk
10001 gfhgdgh
10002 hjkdghkfh
Answer
#1
create an output port and write the expression to replace
values or create a stored procedure and call it expression
transformation
1
Answer
#2
Through Dynamic lookup you can handle it very easily.
Answer
#4
First make sure the values are sorted by rowid and passed
to an expression transformaion.Now create a variable port
in the exp. Lets think var1 and make it empty in the
expression .
IIF(VAR1=ROWID,XX||SUBSTR(ROWID,3,3),ROWID)
Thanks
Abhishek
Answer
#1
Use delete * .... in pre-sql
Answer
#2
Abhishek Bhai,
Possible Solution:-
Cheers !
2
Answer
#3
in the warehouse designer,in targets ,in generate and
execute option truncate the table
Answer
#4
If the informatia user id does not have permissions to
truncate the table in the database. Then there is no other
way we can truncate the table.
Answer
#2
if it is flat file ur answer is write,if it is relational
souce then go to source qualifier properties there u write
the query like
select distinct a.* from t1 a where 10=(select sal from
t1 b where a.sal>b.sal)
i think it is working
Answer
#3
only use the Aggregator function.....
first(sal>=values)
Answer
#4
SQL:
SELECT id, salary from <table name> where rownum <= 10
3
ORDER BY salary DESC;
Answer
#5
it can achieve with inline view query in sqt/r
Answer
#6
use sorter--> expression-->filter
1)sorter descend
2)use sequence generator connected expression to generator
sequence,
3)filter the value sequence number greater than 10
Answer
#7
I hope the simpliest way is what Raka #4 suggested
We can override in the SQL override in the source qualifier
transformation.
Answer
#8
First use sorter with salary based and sequent generator next filter
transformation
--------->Filter(seq<=10)
Answer
#9
Use source Qualifier Transformation,we can edit the default
sql query and write like as
select * from ( select * from emp
order by sal desc)
where rownum <11;
Answer
# 10
do following steps
Answer
#2
They can not be unshared.
5
Because it is to be assumed that users have created
shortcuts to objects in these folders. Un-sharing them would
render these shortcuts useless and could have disastrous
consequences.
Question
How can the following be achieved in 1 single Informatica
Mapping.
===================================================
HEADER
COL1 COL2 COL3 COL5 COL6
1 ABC NULL NULL CITY1
2 XYZ 456 TUBE CITY2
3 GTD 564 PIN CITY3
SUBHEAD
COL1 COL2 COL3 COL5 COL6
1 1001 VAL3 748 543
1 1002 VAL4 33 22
1 1003 VAL6 23 11
2 2001 AAP1 334 443
2 2002 AAP2 44 22
3 3001 RAD2 NULL 33
3 3002 RAD3 NULL 234
3 3003 RAD4 83 31
DETAIL
COL1 COL2 COL3 COL5 COL6
1 D001 TXX2 748 543
1 D002 TXX3 33 22
1 D003 TXX4 23 11
2 D001 PXX2 56 224
2 D002 PXX3 666 332
========================================================
TARGET1
2 XYZ 456 TUBE CITY2
TARGET2
2 2001 AAP1 334 443
2 2002 AAP2 44 22
6
TARGET3
2 D001 PXX2 56 224
2 D002 PXX3 666 332
Answer
#1
Hi,
This could be implemented in many ways.One such way is :
INNER JOIN
DETAIL D
ON
H.COL1=D.COL1
AND
D.COL2 IS NOT NULL AND D.COL3 IS NOT NULL AND D.COL5 IS NOT
NULL AND D.COL6 IS NOT NULL
i.e the above query will fetch the value 2 .Insert this to
a flat file.
Ahmed
7
Answer
#3
1)In the mapping join all the 3 flat files on the common
column(Col1).(2joiners are needed)
2)Create a expression drag all the columns from the end
Joiner to Exp.In the expression create a 3 ports
H_flag,S_Flag,D_flag.
H_Flag =iff(isnull(col2),'x',iff(isnull(col3),'x',iff(isnull
(col5),'x',iff(isnull(col6),'x','y'))
the result of this is if you have null in any header column
this flag value will be 'X'.
Repeat this with Detail and subhead as well.
Declare these flag ports as last ports of expression.
3)Create a router with a group in it for
H_Flag!='x'
S_flag!='x'
D_Flag!='x'
This will get you not null columns from all the three and
accordingly u can load the target.
But if you have any other conditions ,still you can use
these flags in filters and control the data flow.
Question
how to join the two flatfiles using the joiner t/r if there
is no matching port?
Answer
#1
yes you can join two flatfiles using the joiner t/r even if
you don't have any matching port for that you need to take
one dummy column in source files and based on the dummy
column you can join them
Answer
#2
Hi Jana,
How can we take a dummy column i.e what r the values would
be in that port? We need to take dummy columns on both files?
8
In the lookup condition use dummy1=dummy2. This is going to be always true so all
the records are took into consideration.
Question
What is the difference between Oracle performance and
Informatica Percfomance? Which performance is better?
Answer
#1
Generally we check for busy percentage in session log if
its source busy % or Tgt% is more then we try to do
database tunning,if its transformance % is more then go for
Informatica part.But form my side I would rather suggest
for both kind of performance
Answer
#2
oracle performence deals with the source &targets.
informatica performanc deals with the tranformations.
for effficient result both are impotent...
Answer
#3
Orcale performance is better.Because informatica is nothing
but a metadata.Informatica is there only to play with data.
Ya but at the end of the day what matters is how you design
the logic.
Question
How to send duplicates to one target and unique rows to one
target?target is empty
Answer
#1
1> using dynamiclookup concept
2> using variable concept
First solution
source > sorter >dynamic lookup > filter > Target1 and
Target2
Answer
#2
source> dynamic lookup>router,2 conditions 1. condition
if column_lkp port is null then insert into target1(unqie)
2. condtion if COLumn_lkp port is not null then insert into
target2(duplicates)
Answer
#4
we can do this process by 2 ways ....
1)by dynamic lookup option in lookup(new lookup row) we can
load duplicate rows in one target table and unique rows in
one target table
do to this we need to have router transformation (add to
group ports one is for unique(new lookup row=1) and other is
for duplicate(new lookup row =2))after the lookup trans.
2)we can perform this by aggregator transformation using
coutnt(*) >1 for duplicate rows .here also we need to use
9
router transformion.
Answer
#5
Using Source Qualifier Trnsformation ,
Explantion:
1.Take 2 Source Qualifier Transformations,and
2.One sq ports connect to Target(Unique Target) then Write
a SQL Query (sqlOverride) ,
SELECT DISTINCT EMPNO,ENAME
FROM EMP;
3.TAKE ANOTHER SQ AND CONNECT TO ALL PORTS TO TARGET,THEN
DEVELOP THE SQLOVERRIDE,
SELECT * FROM EMP WHERE ROWID IN(SELECT ROWID FROM EMP
MINUS
SELECT MAX(ROWID) FROM EMP
GROUP BY EMPNO,ENAME)
Question
How to load relational source into file target?
Answer
#1
use mapping designer create source db file as well as tgt
flat file with required column information .
Now you need to pull this record when it was added as well
as when it was modified to keep your warehouse in sync with
OLTP system.
13
Ho to handle changing source file counts in a mapping?
Answer
#1
we can maintain versions depends upon changes of sourece
file requirement
new_seq=max(mapping_variable+sequence_generator (value)
then it is like 2million+1...2......(seq_gen)
16
alternatively you could use unix commanda in the command
task.
Answer
#2
You can use a pmcmd command to trigger the workflow using a
shell script. You can also use the event wait task
described in the above answer.
17
Question
what are testing in a mapping level please give brif
eplanation
Answer
#1
Hi,
18
pipeline is the collection fo sources, transformations and
targets that receive data from a single active source.
Question
which T/r we can use it mapping parmeter and mapping
variable? and which one is reusable for any mapping mapping
parmeter or mapping varibale?
Answer
#1
i believe it is seq gen, filter, expression in which v can
use mapping parameter and variable.
19
result of an expession.
reach me on 9866188658.
Answer
#2
By using Update override option in Target table we can
update the table
Answer
#3
We can use update override in the target table in mapping.
This would automate the process and you will not ahve to do
it manually
Question
how can u connect client to ur informatica sever iff server
is located at different place( not local to the client)
Answer
#1
Through IP Address
Answer
#2
Hi U need to connect remotly to ur server and access the
repository.
U will be given repository user name and pwd and add this
repositiory and connect to it with ur credentials.
Answer
#3
you to connect to server
Question
wt is informatica file watch timers
in aflat i want to get the first record and last record how
could i.
Answer
#1
i believe this should work.
Answer
#7
It may be ETL Tools like Informatica,Datastage,etc.,
Answer
#8
Data Object: intermediate table if we r using or Stage
table (temp table) which are the resulting data object
after applying the transformation.
Answer
#9
23
The main data object present inbetween source and target is
staging layer only, Staging layer will do eliminate the
inconsistency data and gives the result data object
Answer
# 11
source qualifier is the correct answer,because with out
source qualifier u cant do any thing
Answer
# 12
Answer is Source Qualifier.
24
repositories , but we can not share the matadata accross
the repositories,
25
Informatica has own Sheduling components .Since Informatica
widely used as Data integration and Data warehousing Entity
and there are lots of depedencies on other jobs ,reason
informat application is scheduled by external sheduluing
device such as
1) Control M
2) Crontab
3) maestro
c1 s1 c1
c1 s2 s1
c1 s1 c1
c2 s3 s2
c3 s4 c1
c3 s2 s1
c2
s3
c3
s4
c3
s2
you can drag that target Warehouse desiner ,and mark those
columns which are no keys as a key for time being ,on whihc
you are trying to update. even though those columns are
not keys attributes or key columns at database level,for
time being those will be treated as key elements, hense you
can apply or use update strategy .
Pass VAR2 and VAR3 to the router. Have one output group
with condition IS_NUMERIC(VAR2) and the other obviously is
the default group. For the first group connect VAR2 to ID
of target and VAR3 to NAME of target. For default connect
VAR2 to NAME and VAR3 to ID
Output
=====
ID NAME
= =====
1 a
1 a
Answer
#2
Here you should not use a router as it sends the data to
two different target or two instances of the same target.
29
Link the columns to the target now. Done!
Question
what is metadata?
Answer
#1
Metadata is data about data..it will have all information
about mappings and transformations.
nswer
#3
metadata is a structure of data, information about
mappings and transformations
Answer
#5
Commonly known as "data about data" it is the data
describing context, content and structure of records and
their management through time
Answer
#6
A metadata is Data about data . The repositary contains the
metadata , it means all the information of the mappings,
task.....etc
Answer
#7
Metadata is nothing but data about data.As metadata
contains information related to data from where data is
and connections to data.
Question
Explain about HLD and LLD ?
Answer
#1
HLD refers High Level Design and
LLD refers Low Level Design
LLD: It reveals the complete details to be complted to get the product finished.
It gives the designing of inside modules of blocks specified in HLD.
For example, HLD is just representing a computer by different blocks like CPU, I/O
devices, Memory etc. Where as LLD with respect to this example is detailed
description of all the blocks of HLD like CPU, I/O devices, Memory etc.
Question
for ex: in source 10 records are there with column sal. use
a filter transformation condition as Sal=TRUE and connect
to target. what will happen.
31
Answer
#1
I checked result.
Steps:-
33
2.generate the target tables.
Note:-
send your responses to suriaslesha_sreekar@yahoo.co.in
Answer
#2
Can it be done in this way....
Thanks
Anand Kumar
Answer
#4
1. drag source and targets in to mapping designer work space
2. from t/r devoloper take sequence genarato t/r,exp t/r
and router t/r.
34
3. in seqg t/r give startvalue 1 ,increment by 1
give netval to a newpor in exp t/r
4. drag all ports of sq to exp in addition to newport
5. in router t/r create one group name as odd
give condetion mode(newport/2)!=0
6. give from group odd to t1
and defoult to t2.
Question
why cant we put a sequence generator or upd strategy
transformation before joiner transformation?
Answer
#1
Joiner is to join two different sources .If u use Update
strategy T.F and trying to use DD_delete&DD_reject option
the some of the data will get deleted and u can't see the
at joiner output.
So we can't go for this .
Question
In Real Time what are the scenarios u faced, what r the
tough situations u have overcome, and explain about sessions.
Answer
#1
Getting first job in Informatica and working with
transformations...
Question
how u know when to use a static cache and dynamic cache in
lookup transformation.
Answer
#1
if you need the source data after look up transformation to
transfer any other transformation more than once you can use
dynamic cache otherwise use static cache
Answer
#2
Dynamic cache is generally used when u are applying lookup
on a target table and in one flow same data is coming twice
for insertion or once for insertion and once for updation.
Performance: dynamic cache decreases the performance in
comparision to static cache since it first looks in the
whole table tht whether data ws previously present if no
thn only it inserts so it takes much tym
Static cache do not see such things just insert data as
many tyms as it is coming.
Answer
#3
when we use conneted type of lookup then we use dynamic
caches& when we use unconncted lookup in the modelling we
use static cache
Answer
#5
hi neetu,
35
i hope you were a bit wrong.
static cache is also used to check whether row exists or
not.....
well krishna for your question on dynamic cache..
static cache is one in which data when changed in the
target table cannot be incorporated in the look up on the
fly.
when you use a dynamic cache any changes to the target
table gets reflected in the lookup even if the record is
being inserted or updated in the same mapping run priorly
Answer
#6
By default lookup is static.
Dynamic is used whn the records coming frm the source to
target in multiple times. i.e if
1)
emp_id=101...city=hyd...age=25...
(first this is inserted to target)
2)at second time when 101 employee is changing his city frm
hyd to chennai here we need to update the target table with
city name as chennai. thn how lookup cache knows tht a
record is updated...
3) if it is static lookup, after updating tht record it
will not refresh the lookup cache.... where as if it is
dynamic cache it will refresh the cache....
4) Based on our requirement we r going to use.....
Question
whether Sequence generater T/r uses Caches? then what type
of Cache it is
Answer
#1
multi-caches.
Answer
#2
no it won't have any cache
we have caches for the following t/r
aggregate t/r
joiner t/r
sorter t/r
lookup t/r
Answer
#3
also a rank Xmation uses cache
Answer
#5
the seq t/r uses index cache for the sequential range of
numbers for the generated keys.
Answer
#6
Sequence generator uses a cache when reusable.
37
are product, store, customer, sales ticket, and time.
At a bank ATM, a fact table record is created for every
customer transaction. The dimensions of this fact table
record are financial service, ATM location, customer,
transaction type, and time.
When the telephone rings, the phone company creates a fact
table record for each "hook event." A complete call-
tracking data warehouse in a telephone company records each
completed call, busy signal, wrong number, and partially
dialed call.
In all three of these cases, a physical event takes place,
and the data warehouse responds by storing a fact table
record. However, the physical events and the corresponding
fact table records are more interesting than simply storing
a small piece of rev enue. Each event represents a
conscious decision by the customer to use the product or
the service. A good marketing person is fascinated by these
events. Why did the customer choose to buy the product or
use the service at that exact moment? If we only had a
dimension called "Why Did The Customer Buy My Product Just
Now?" our data warehouses could answer almost any marketing
question. We call a dimension like this a "causal"
dimension, because it explains what caused the event.
Answer
#3
i sincerely appreciate ur interest n moreover ur patience
in explaining so beautiful. thanks a million for ur ans.
can i have your mail id or number pls.
here are mine.
Question
Without using any transformations how u can load the data into
target?
Answer
#1
Simply connected souce with target.
38
Answer
#7
Hi All,
4) event wait will wait for touch file, once it get the
touch file, it will start the load.
Question
explain the scenario for bulk loading and the normal
loading option in Informatica Work flow manager ???
Answer
#1
1)Bulkload & Narmal load
Example::
Students Attendtance... It consists only student
information nothing other than present or absent..i.e
Boolean values(Yes or No)
Answer
#5
A FACT TABLE WITHOUT FACT IS CALLED FACT LESS FACT TABLE
SUPPOSE WE NEED TO COMBINE TWO DATAMARTS ,ONE DATAMART
CONATINS FACT LESS FACT TABLE AND ANOTHER DATAMART WITH
43
THE FACT TABLE
FACT LESS FACT TABLES ARE USED TO CAPTURE DATE TRANSACTION
EVENTS
Answer
#6
The fact table which contains the business events or
coverage that could be represented in a fact table, but
there will be no measures or facts associated with these.
Question
In real time scenario where can we use mapping parameters
and variables?
Answer
#1
Mapping & Mapplet, v can use the mapping parameter &
variable. V can also create the parameter & variable in the
sesssion level.
Answer
#2
Before using mapping parameters and mapping variables we
should declare these things in mapping tab of mapping
designer.
Example:::
if we declare mapping parameter we can use that parameter
untill completing the session,but if we declare mapping
variable we can change in between sessions.Use mapping
variable in Transcation Control Transformation......
Answer
#3
mapping variable unlike mapping parameter changs its value
during session execution. it is used in incremental loading.
44
write opposite condition what mentioned in filter
transformation1 in another filter t.f and pass the rows
which are not satisfying filter1 but satisy 2 in to target2
Answer
#3
Don't use filter transformation.
iif is_null(column_name,-NA-,column_name)
Question
ONE FLAT FILE IS THERE WHICH IS COMMA DALAMETED . HOW TO
CHANGE THAT COMMA DELEMITER TO ANY OTHER AT THE TIME OF
RUNNING ?
Answer
#1
I think we can change it in session properties of mapping
tab.if select flatfile on top of that we see set file
properties.
Question
two tables from two different databases r there . both
having same structure but different data . how to compare
these two tables ?
Answer
#1
Using Joiner Transformation, to compare the two tables.
Answer
#2
That is not a correct answer because joiner is used for
joini two tables
so at source analyser right click on sourc table click on
compare by this u can get
this is i have seen some where
Answer
#4
if ur comparisiom means joining it shud be using joiner
t/tion using a join condition and join type
Answer
#9
Joiner transfomation is used to join two differned soures
from the same database. Lookup transformation and
expressions(if needed) can be used to compare data from two
different two differebt types of sources.
46
Answer
# 10
If u want to compare the data present in the tables go for
joining and comparison..
if u want to compare the metadata (properties)of the tables
to for "compare Objects" in source analyser/
Question
IN SCD TYPE 1 WHAT IS THE ALTERNATIVE TO THAT LOOKUP
TRANSFORMATION ?
Answer
#1
Alternative to Lookup id Joiner.we need to import source
structure of target in Source Analyser and bring that into
mapping and use it for comparison like Lookup.
Answer
#4
You can use Joiner transformation to design scd Type1
manually. Import target as source and use joiner
transformation. Use expression to insert and update the
rows into target.
Question
1)can anyone explain how to use Normalizer transformation
for the following scenario
2. Sub_id Subname
------- -------
10 ENG
47
20 MAT
30 ART
Answer
#3
It is useful in combining all the multiple columns in to a
single column and viceversa.
Let me know if this is wrong.....!
Question
In a mapping i have three dimensions. If i want to pass a
same surrogate key value to all the three dimensions by
using one sequence generator is possible?If the mapping is
containing single flow? And in the same case if the mapping
is contaning 3 flows for the three dimensions then by using
one sequence generator can we populate surrogate key (same
value) to all the three dimensions?
Answer
#1
u can pass same surrogate key to three dimensions ,this is
possible in case 2 i.e 3 flows,according to my view
Answer
#2
yes we can pass same surrogate key for 3 dimentions. bcz
three dimention are involved in same mapping. we can also
reuse the sequence generator.
Answer
#3
Hi i have a samll doubt on this
Three dimensions are in same mapping but three different
flows. 3 diffrent flows means when the first completes then
only the second flow will start in this case by use one
sequence generator how will we pass the same values to all
the three dimensions. The surrogate key value should pass
like this for all the dimensions
48
dim1 dim2 dim3
---- ---- -----
1 1 1
2 2 2
3 3 3
4 4 4
Answer
#4
Use the Sequence and Expresion transfermations.first
genarate the surrogate with Seq trans,then send values to
exp trans,connect the exp trans o/p ports to 3 dimentions.
49
not in sql over ride(According to my knowledge)
50
Answer
#7
Filter, Expression & update strategy
Answer
#8
u can use any T/r based on Exp editor available in which
t.r that t/r u can use .lkp value.u can use
update ,agg,rank,filter,sq
Question
what r the transformations that r not involved in mapplet?
Answer
#1
1.Normaliaer
2.xml source qualifier
3.target definions
4.sequence generator
Answer
#3
*You cannot include the following objects in a mapplet:
1.Normalizer transformations
2.COBOL sources
3.XML Source Qualifier transformations
4.XML sources
5.Target definitions
6.Other mapplets
7.Pre- and post- session stored procedures
Answer
#6
You cannot include the following objects in a mapplet:
- Normalizer transformations
- Cobol sources
- XML Source Qualifier transformations
- XML sources
- Target definitions
- Pre- and post- session stored procedures
- Other mapplets
Question
what is the function of 'F10' informatica ?
Answer
#1
used in debugging process
Answer
#2
F10 and F5 are used in debugging process
Question
What is data merging, data cleansing and sampling?
Answer
#1
Data Cleansing: A two step process of detection and
correction of errors in a data set.
Answer
#2
data merging :multiple detailes values are summarised into
single summaeised value.
data cleansing:to eliminate the inconsistant data
sampling:it is the process ,orbitarly reading the data from
group of records.
53
Answer
#4
datacleaging:it is the process of identifying and changing
inconsistency and inacquries
datamerging:it is process of integreated multiple
inputsource into singleoutput with similar srtucture and
datatype
Answer
#5
The main thing Merging of data is nothing but integrating from multiple source
systems. It is in 2 types
1.Horizontal merging(Join)
2.Vertical Merging(Union)
Question
What are the different options used to configure the
sequential batches?
Answer
#1
Two options
55
5. we cant use cobol source qualifier,joiner,normalizer
transformations in mapplet.
Answer
#5
MAPPLET-REUSABLE COLLETION OF TRANSFORMATION CREATED TO
SOLVE A LOGIC.
58
nswer
#1
type2 scd it wil maintain historical informtion + currnt
information along with 3 options .....
1.effective date
2.version number
3.flag value
Question
Define informatica repository?
Answer
#1
Hi
Question
WHAT IS THE NAME OF THAT PORT IN DYNAMIC CACHE WHICH IS USED
FOR INSERT , UPDATE OPRATION ?
Answer
#1
Associate port
Answer
#3
u mean inserting and updating the source records into the
cache. then the answer shud be insert else update which
lies below dynamic lookup port.
59