Why is meant by direct and indirect loading options in sessions?
Answers were Sorted based on User's Feedback
Answer / raj
we use file type direct when we are loading single file into
target. we use Indirect when we want to load multiple files
through single session.
| Is This Answer Correct ? | 21 Yes | 1 No |
Answer / prashant
When we are using the Direct option at the session
level,then we are loading only single file.
But when we are using Indirect option, then we can load
multiple files also.For this we have give a list file
(.list) path in the session which will contain the names of
the files need to be loaded.
| Is This Answer Correct ? | 15 Yes | 1 No |
Answer / pradeep
we use file type direct when we are loading single file into
target. we use Indirect when we want to load multiple files
through single session in the mapping
| Is This Answer Correct ? | 7 Yes | 1 No |
Answer / srikanth chowdary
when you are using the direct method option in session side that time your loading only single file into target and when you are using indirect method that time you can create LIST.txt file then you can pass multiple files into target in a single session but need to same structure in files
| Is This Answer Correct ? | 1 Yes | 0 No |
What are the methods for creating reusable transforamtions?
HOW TO DELETE A RECORD FROM FLAT FILE
Source-1 No name 1 satish 2 karthik 3 swathi 4 keerthi Source-2 No name 1 satish 2 karthik 5 santhose 6 vasu Target 3 swathi 4 keerthi 5 santhose 6 vasu here i want non matching Records i want how to achieve that
If we are using an aggregator but forget to mention the group by port .what will be the output??
One of the optimizing technique to improve the session performance is push down optimization,by using push down optimization we push as much as transformation logic to source/target database,but this degrades the d/b performance,how to overcome this?
write a query to follwoing table amount year quarter 254556 2003 first 546332 2003 second 129034 2003 third 490223 2003 fourth 165768 2004 first 265443 2004 second 510412 2004 third 435690 2004 fourth i want the output year q1_amount q2_amount q3_amount q4_amount 2003 254556 546332 129034 490223 2004 165768 265443 510412 435690
what is the complex transformation you used in informatica
There are 2 files, Master and User. We need to compare 2 files and prepare a output log file which lists out missing Rolename for each UserName between Master and User file. Please find the sample data- MASTER.csv ---------- Org|Tmp_UsrID|ShortMark|Rolename ---|---------|----------|------------ AUS|0_ABC_PW |ABC PW |ABC Admin PW AUS|0_ABC_PW |ABC PW |MT Deny all GBR|0_EDT_SEC|CR Edit |Editor GBR|0_EDT_SEC|CR Edit |SEC MT103 GBR|0_EDT_SEC|CR Edit |AB User USER.csv -------- Org|UserName|ShortMark|Rolename ---|--------|---------|------------ AUS|charls |ABC PW |ABC Admin PW AUS|amudha |ABC PW |MT Deny all GBR|sandya |CR Edit |Editor GBR|sandya |CR Edit |SEC MT103 GBR|sandya |CR Edit |AB User GBR|sarkar |CR Edit |Editor GBR|sarkar |CR Edit |SEC MT103 Required Output file: --------------------- Org|Tmp_UsrID|UserName|Rolename |Code ---|---------|--------|------------|-------- AUS|0_ABC_PW |charls |ABC Admin PW|MATCH AUS|0_ABC_PW |charls |MT Deny all |MISSING AUS|0_ABC_PW |amudha |ABC Admin PW|MISSING AUS|0_ABC_PW |amudha |MT Deny all |MATCH GBR|0_EDT_SEC|sandya |Editor |MATCH GBR|0_EDT_SEC|sandya |SEC MT103 |MATCH GBR|0_EDT_SEC|sandya |AB User |MATCH GBR|0_EDT_SEC|sarkar |Editor |MATCH GBR|0_EDT_SEC|sarkar |SEC MT103 |MATCH GBR|0_EDT_SEC|sarkar |AB User |MISSING Both the files are mapped through Organization, Shor_mark. So, based on each Organization, Short_Mark, for each UserName from User.csv, we need to find the Matching and Missing Rolename. I am able to bring Matching records in the output. But really I don't find any concept or logic to achieve "MISSING" records which are present in Master and not in User.csv for each UserName. Please help out guys. Let me know if you need any more information. Note:- In User.csv file, there are n number of Organization, under which n number Shortmark comes which has n number of UserName.
Suppose i have 10000 records.First time i have to load 1 to 1000 records and second run i have load 1000 to 2000 records and third load i have to load 2000 to 3000 rows .How will achieve
How do you load alternate records into different tables through mapping flow?
daily how much amount of data send to production?
Hi ETL gurus can any one tell me with a flow how to implement SCD Type 1 and SCD Type 2 in a single mapping.For some fields SCD type has to be implemented and for some fields scd type has to be implementd..Thank in advance..please let me know in case of any concerns...