Golgappa.net | Golgappa.org | BagIndia.net | BodyIndia.Com | CabIndia.net | CarsBikes.net | CarsBikes.org | CashIndia.net | ConsumerIndia.net | CookingIndia.net | DataIndia.net | DealIndia.net | EmailIndia.net | FirstTablet.com | FirstTourist.com | ForsaleIndia.net | IndiaBody.Com | IndiaCab.net | IndiaCash.net | IndiaModel.net | KidForum.net | OfficeIndia.net | PaysIndia.com | RestaurantIndia.net | RestaurantsIndia.net | SaleForum.net | SellForum.net | SoldIndia.com | StarIndia.net | TomatoCab.com | TomatoCabs.com | TownIndia.com
Interested to Buy Any Domain ? << Click Here >> for more details...


What is configuration your file structure 2)I have two
databases both are Oracle while loading data from source to
target the job takes 30 min but I want to load less time how?



What is configuration your file structure 2)I have two databases both are Oracle while loading data..

Answer / subhash

1) configuration file structure:
{
node "node1"
{
fastname "<SERVER_NAME>"
pools ""
resource disk "</dsg/IBM/..DATASETS Path>" {pools
""}
resource scratchdisk "</dsg/..Scratch disk/Buffer
Path>" {pools ""}
}
node "node2"
{
fastname "mctux034"
pools ""
resource disk
"/dsg/IBM/IBM/InformationServer/Server/Datasets" {pools ""}
resource scratchdisk
"/dsg/IBM/IBM/InformationServer/Server/Scratch" {pools ""}
}
}
2)
SRC--->COPY---->TGT
Here Copy stage increase performance and works as a Buffer
when diff in reading count and writing record count.
by setting/checking below in Target ORACLE Connector:
Record Count=6000 (default is 2000, should be multiple of
Array Size)
Array Size=3000 (default is 2000)
Write mode=Bulk Load
Rebuild Indexes After Bulk Load

Is This Answer Correct ?    3 Yes 0 No

Post New Answer

More Data Stage Interview Questions

Instead of using shared container in a job, I use jobs which perform similar function as Container in the sequence. Then what is the need of Shared Container?

1 Answers   TCS,


what is sparse lookup?

7 Answers   IBM,


What are the difference types of stages?

0 Answers  


i have a some records eno ename acono amount 1001 suresh sbi101 12000 1001 suresh sbi101 14000 1001 is dublicata how can i remove duplicate?

4 Answers  


I have source file which contains duplicate data,my requirement is unique data should pass to one file and duplicate data should pass another file how?

7 Answers   CTS,


Can anyone tell me a difficult situation who have handled while creating Datastage jobs?

0 Answers   Cap Gemini, TCS,


What all are the different way to run a job?

0 Answers  


what is repositery?

0 Answers   IBM,


How many types of views are there in a datastage director?

0 Answers  


What is Cleanup Resources and when do you use it?

2 Answers  


What are the different options associated with dsjob command?

0 Answers  


how to delete one row in target dataset

0 Answers   Satyam,


Categories