HOW CAN WE SEE THE DATA IN DATASET?
Answers were Sorted based on User's Feedback
Answer / bharathd
Thanks sekhar for Nice info...
We can also view the data set through datastage client
called designer, for this go to tools menu->clik on dataset
management->select the dataset whichnone you want to view-
>the click ok->click on show datawindow(in cube shape)-
>click ok .......
| Is This Answer Correct ? | 6 Yes | 0 No |
1--> Way through DataStage:
through datastage client - designer, for this go to
tools menu->
clik on dataset management ->
select the dataset which one you want to view ->
then click ok ->
click on show datawindow(in cube shape)->
click ok .......
2--> Way through UNIX:
1. Before using orchadmin, you should make sure that either the working directory or the $APT_ORCHHOME/etc contains the file “config.apt” OR The environment variable $APT_CONFIG_FILE should be defined for your session.
Orchadmin commands
1. CHECK: $orchadmin check
Validates the configuration file contents like , accesibility of all nodes defined in the configuration file, scratch disk definitions and accesibility of all the nodes etc. Throws an error when config file is not found or not defined properly
2. COPY : $orchadmin copy <source.ds> <destination.ds>
Makes a complete copy of the datasets of source with new destination descriptor file name. Please not that
a. You cannot use UNIX cp command as it justs copies the config file to a new name. The data is not copied.
b. The new datasets will be arranged in the form of the config file that is in use but not according to the old confing file that was in use with the source.
3. DELETE : $orchadmin < delete | del | rm > [-f | -x] descriptorfiles….
The unix rm utility cannot be used to delete the datasets. The orchadmin delete or rm command should be used to delete one or more persistent data sets.
-f options makes a force delete. If some nodes are not accesible then -f forces to delete the dataset partitions from accessible nodes and leave the other partitions in inaccesible nodes as orphans.
-x forces to use the current config file to be used while deleting than the one stored in data set.
4. DESCRIBE: $orchadmin describe [options] descriptorfile.ds
This is the single most important command.
1. Without any option lists the no.of.partitions, no.of.segments, valid segments, and preserve partitioning flag details of the persistent dataset.
-c : Print the configuration file that is written in the dataset if any
-p: Lists down the partition level information.
-f: Lists down the file level information in each partition
-e: List down the segment level information .
-s: List down the meta-data schema of the information.
-v: Lists all segemnts , valid or otherwise
-l : Long listing. Equivalent to -f -p -s -v -e
5. DUMP: $orchadmin dump [options] descriptorfile.ds
The dump command is used to dump(extract) the records from the dataset.
Without any options the dump command lists down all the records starting from first record from first partition till last record in last partition.
-delim ‘<string>’ : Uses the given string as delimtor for fields instead of space.
-field <name> : Lists only the given field instead of all fields.
-name : List all the values preceded by field name and a colon
-n numrecs : List only the given number of records per partition.
-p period(N) : Lists every Nth record from each partition starting from first record.
-skip N: Skip the first N records from each partition.
-x : Use the current system configuration file rather than the one stored in dataset.
6. TRUNCATE: $orchadmin truncate [options] descriptorfile.ds
Without options deletes all the data(ie Segments) from the dataset.
-f: Uses force truncate. Truncate accessible segments and leave the inaccesible ones.
-x: Uses current system config file rather than the default one stored in the dataset.
-n N: Leaves the first N segments in each partition and truncates the remaining.
7. HELP: $orchadmin -help OR $orchadmin <command> -help
Help manual about the usage of orchadmin or orchadmin commands.
| Is This Answer Correct ? | 2 Yes | 0 No |
Answer / sekhar naskar
If we want to see all the data it is :
orchadmin dump <dataset_name>
If we want to see specific field data it is :
orchadmin dump -name -field <field_name> -field
<field_name> <dataset_name>
But before that the DS environment should be set.
Like below ( but may change based on the configuration):
export PATH
. /opt/IBM/InformationServer/Server/DSEngine/dsenv
LD_LIBRARY_PATH=$APT_ORCHHOME/lib:$LD_LIBRARY_PATH; export
LD_LIBRARY_PATH
| Is This Answer Correct ? | 1 Yes | 0 No |
how to connect source to db?generally what r stages u used? how to find the data is having delimiter format?
Scenario : I have 2 jobs say job A and Job B with parameters x and y respectively. I need to create a sequence job. If we pass parameter x then Job A should run, If we pass parameter y then Job B should run, if we dont pass any parameter then Both Job A & B should run.
What is meta stage?
create a job that splits the data in the Jobs.txt file into four output files. You will direct the data to the different output files using constraints. • Job name: JobLevels • Source file: Jobs.txt • Target file 1: LowLevelJobs.txt − min_lvl between 0 and 25 inclusive. − Same column types and headings as Jobs.txt. − Include column names in the first line of the output file. − Job description column should be preceded by the string “Job Title:” and embedded within square brackets. For example, if the job description is “Designer”, the derived value is: “Job Title: [Designer]”. • Target file 2: MidLevelJobs.txt − min_lvl between 26 and 100 inclusive. − Same format and derivations as Target file 1. • Target file 3: HighLevelJobs.txt − min_lvl between 101 and 500 inclusive. − Same format and derivations as Target file 1. • Rejects file: JobRejects.txt − min_lvl is out of range, i.e., below 0 or above 500. − This file has only two columns: job_id and reject_desc. − reject_desc is a variable-length text field, maximum length 100. It should contain a string of the form: “Level out of range: <min_lvl>”, where <min_lvl> is the value in the min_lvl field. My Question is how do you write the stage variable for reject rows.
in datastage scenario this is sourec data i want this traget? tell me please? source target ------- ------- Mohan,Prasad,Naresh Mohan Prasad Naresh
Anyone has Datastage certification free dumps for 000-418 , 000-421 codes, mail me @ manik.dwh@gmail.com 000-418 : InfoSphere DataStage v8.0 000-421 : InfoSphere DataStage v8.5
what will happen if we allow duplicates in datastage lookup abort drop record 1st value of duplicate record none
How you can fix the truncated data error in datastage?
Is there any possibility to generate alphanumeric surrogate key?
In the source seq file have 2 columns, i.e input: col1,col2 1,1 2,rajesh 3,15000 4,2 5,suresh 6,16000 7,3 8,veeru 9,17000 I want in the output like: eno,ename,sal 1,rajesh,15000 2,suresh,16000 3,veeru,17000 any one answer this scenario
Hi This is Vijay, How can your remove the duplicates in sequential File?
15 Answers HCL, Scope International,
Hi dude, i/p eno ename date edate deptno 1 x 10thjuly 30july A 2 y 10th aug 30 aug B 1 x 1aug ------ A 3 z 1jan ------ c 2 y 1sep ------ B o/p is eno ename date edate deptno 1 x 1aug ------ A 2 y 1sep ------ B 3 z 1jan ------ c Here edate is System generates we don't know that value. Please design the job by using SCD stage and Sql query. If any body know this answer please tel me. Tanks.