3,if our source containing 1 terabyte data so while loading
data into target what are the thing we keep in mind?
Answers were Sorted based on User's Feedback
Answer / chinni
1tera data is huge amount of data so if we use normal load
type it takes so much of time....to overcome this better to
use bulk loading...and also taking the large commit
intervals we can get good performance....and one more
technique is external loading.....
| Is This Answer Correct ? | 9 Yes | 1 No |
I agree with you Chinni.
But my suggestion is instead of loading huge amount of data
at a time,we can always split the source if it is a file
using unix and extract the data from all the source files
using indirect type and load into target.
Please let me know if you have any suggestions.
Thanks.
James.
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / chiranjeevi k
We can use push down optimizationto to handle the huge
volume of data.
Thanks
Chiranjeevi
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / bakshu shaik
one more thing we need to take into consideration, along with above suggestions:
If the target having an index, Loading time (huge amount of data) increases/performance hazard......so it's always better to drop the index before loading(Stored procedure transformation) and create again after loading completed...!
| Is This Answer Correct ? | 0 Yes | 0 No |
Along with the above suggestions......
Better to use partitioning (single or multiple thread) and and increase the commit level to 50 lacs. Better to load first flat file instead of loading into table.
then if u r db is oracle better to use sql loader to load data from file to table.If DB2 than use db2 loader.
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / upendra
large amount of data in source file.So we can devide partitions and commit level ur own wish,then connected to different target table, each one target table better for bulk load.
| Is This Answer Correct ? | 0 Yes | 0 No |
how lookup is acting as both active and passive ? how is it possitble?
what are the real time problems generally come up while doing or running mapping or any transformation?
How can you join two tables without any common column present in them?
how we can load rejected record's at run time?(not through bad files)
how can u tune u r informatica mapppings
Can anyone tell me, when a flat file contains different kinds of delimeters available for each line ( For Ex: First line contains [,*&#$] and second line contains [%@$#%^]. Like that all records having different kinds of delimeters. How can we implement this kind of flat file in informatica?
Why we use ENABLE HIGH PRECISION in session properties?
What is a poling?
without matching columns in two tables. how can you join
Enlist the differences between database and data warehouse.
how will you get 21 to 30 record from 50 records?
how to get the data from the client machine and how to get server location data to client loction can any body explain to me