Golgappa.net | Golgappa.org | BagIndia.net | BodyIndia.Com | CabIndia.net | CarsBikes.net | CarsBikes.org | CashIndia.net | ConsumerIndia.net | CookingIndia.net | DataIndia.net | DealIndia.net | EmailIndia.net | FirstTablet.com | FirstTourist.com | ForsaleIndia.net | IndiaBody.Com | IndiaCab.net | IndiaCash.net | IndiaModel.net | KidForum.net | OfficeIndia.net | PaysIndia.com | RestaurantIndia.net | RestaurantsIndia.net | SaleForum.net | SellForum.net | SoldIndia.com | StarIndia.net | TomatoCab.com | TomatoCabs.com | TownIndia.com
Interested to Buy Any Domain ? << Click Here >> for more details...

I have load a Dataset in UAT with 2 Node configuration, imported the job into PROD environment which is 4 node configuration and using this DataSet as SRC to other job. will the job run fine or give any errors? If job runs fine, on how many nodes? 2 nodes or 4 nodes?

Answer Posted / ankit gosain

Hi,

Job will run smoothly but the dataset will run on 2 node
configuration.
This is because whenever you create a dataset, 4 componets
are created which are Discriptor file, Data file, Header
File and Control file. Now, when that dataset was created
with 2 node configuration, it captured the configuration
file into it's Discriptor file and now if you'll use that
dataset in some other environment with 4 node
configuration, it won't affect anything, since the
discriptor have the configuration of nodes when it was
creadted.
You can cross check it by using the environment variable
APT_DUMP_SCORE.
If you have further doubts or query, you can mail me at
ankitgosain@gmail.com

Cheers,
Ankit

Is This Answer Correct ?    6 Yes 1 No



Post New Answer       View All Answers


Please Help Members By Posting Answers For Below Questions

hi.... am facing typical problem in every interview " I need some critical scenarios faced in real time" plz help me guys

2923


what is ds administrator used for?

1215


Define ds designer?

1324


Define data aggregation?

1145


What is a folder? Difference types of stages?

1236


How complex jobs are implemented in datstage to improve performance?

1091


What are data elements?

1230


How many Key we can define in remove duplicate stage?

1434


What could be a data source system?

1090


What all the types of jobs you developed?

1234


Define meta stage?

1254


What are the difference types of stages?

1284


Define oconv () and iconv () functions in datastage?

1462


create a job that splits the data in the Jobs.txt file into four output files. You will direct the data to the different output files using constraints. • Job name: JobLevels • Source file: Jobs.txt • Target file 1: LowLevelJobs.txt − min_lvl between 0 and 25 inclusive. − Same column types and headings as Jobs.txt. − Include column names in the first line of the output file. − Job description column should be preceded by the string “Job Title:” and embedded within square brackets. For example, if the job description is “Designer”, the derived value is: “Job Title: [Designer]”. • Target file 2: MidLevelJobs.txt − min_lvl between 26 and 100 inclusive. − Same format and derivations as Target file 1. • Target file 3: HighLevelJobs.txt − min_lvl between 101 and 500 inclusive. − Same format and derivations as Target file 1. • Rejects file: JobRejects.txt − min_lvl is out of range, i.e., below 0 or above 500. − This file has only two columns: job_id and reject_desc. − reject_desc is a variable-length text field, maximum length 100. It should contain a string of the form: “Level out of range: ”, where is the value in the min_lvl field. My Question is how do you write the stage variable for reject rows.

2775


what is the use of materialized view?

1629