Why do flat file load is faster if you compare that with
table load ?
Please answer me.
Advance Thanks,
Manojkumar
Answers were Sorted based on User's Feedback
Answer / joe
In case of dbases- time is consumed to make the dbase
connection for reading/writing data to the table.
| Is This Answer Correct ? | 3 Yes | 0 No |
Answer / suresh
Here the records has to pass through the database drivers and ODBC connections are required, which is a time taking process.
But where as records from flat files are copied in easier way
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / nidhi
To write data in table informatica server generate the insert statement for each row where as for flat file write the data row wise in file.
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / ankur saini er.ankur861@gmail.
Hi in case of teradata..IS have to take care of database
logs , and backup tables etc like transient journel etc
which makes it slow in case of table.
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / sanjay
While loading the flat file target integration service do not have to prepare insert/update/delete sql statements, also there should not be any database overhead like unique index validation or constraints checking for the target data and no auto data type conversion takes place while writing to flat file target.
| Is This Answer Correct ? | 0 Yes | 0 No |
Thanks Rayudu.
Is there any other reson why flat file load is faster than
table load.
I had an interview and even I said the same answer. But the
interviewer asked me "Is there any other reason ?" .
Interviewer didn't satisfied with this answer.
I hope there will be some other reason as well for this
question.
Anyways thanks for your reply.
| Is This Answer Correct ? | 0 Yes | 1 No |
Explain is there any way to read the ms excel datas directly into informatica? Like is there any possibilities to take excel file as target?
How to transform normalized data to denormalized form in informatica? Is there any logic or any transformations to achieve this?
how to we create datamart?
I have a scenario with router transformation. but i dont want to capture records from default group. how can we achieve this.
what is row level and column level indicator?
There are 2 files, Master and User. We need to compare 2 files and prepare a output log file which lists out missing Rolename for each UserName between Master and User file. Please find the sample data- MASTER.csv ---------- Org|Tmp_UsrID|ShortMark|Rolename ---|---------|----------|------------ AUS|0_ABC_PW |ABC PW |ABC Admin PW AUS|0_ABC_PW |ABC PW |MT Deny all GBR|0_EDT_SEC|CR Edit |Editor GBR|0_EDT_SEC|CR Edit |SEC MT103 GBR|0_EDT_SEC|CR Edit |AB User USER.csv -------- Org|UserName|ShortMark|Rolename ---|--------|---------|------------ AUS|charls |ABC PW |ABC Admin PW AUS|amudha |ABC PW |MT Deny all GBR|sandya |CR Edit |Editor GBR|sandya |CR Edit |SEC MT103 GBR|sandya |CR Edit |AB User GBR|sarkar |CR Edit |Editor GBR|sarkar |CR Edit |SEC MT103 Required Output file: --------------------- Org|Tmp_UsrID|UserName|Rolename |Code ---|---------|--------|------------|-------- AUS|0_ABC_PW |charls |ABC Admin PW|MATCH AUS|0_ABC_PW |charls |MT Deny all |MISSING AUS|0_ABC_PW |amudha |ABC Admin PW|MISSING AUS|0_ABC_PW |amudha |MT Deny all |MATCH GBR|0_EDT_SEC|sandya |Editor |MATCH GBR|0_EDT_SEC|sandya |SEC MT103 |MATCH GBR|0_EDT_SEC|sandya |AB User |MATCH GBR|0_EDT_SEC|sarkar |Editor |MATCH GBR|0_EDT_SEC|sarkar |SEC MT103 |MATCH GBR|0_EDT_SEC|sarkar |AB User |MISSING Both the files are mapped through Organization, Shor_mark. So, based on each Organization, Short_Mark, for each UserName from User.csv, we need to find the Matching and Missing Rolename. I am able to bring Matching records in the output. But really I don't find any concept or logic to achieve "MISSING" records which are present in Master and not in User.csv for each UserName. Please help out guys. Let me know if you need any more information. Note:- In User.csv file, there are n number of Organization, under which n number Shortmark comes which has n number of UserName.
Define Pmrep command?
What is the difference between a shortcut of an object and copy of an object? Compare them.
Difference between session variables and workflow variables?
What is best approach to load 100 different source files (Different structure) to differet target tables ?
My source is EmpID, Salary- (101, 1000)(102, 2000)(103, 3000). In the Target I want the following EmpID, Salary, Composite_Salary- (101, 1000, 1000)(102, 2000, 3000)(103, 3000, 6000). Please guide on how to build the mapping.
What do you mean by channel change?