How to exclude first and last lines while reading data into a sequential file(having some 1000 records).I guess probably by using unix filter option but not sure which to use
Answers were Sorted based on User's Feedback
Answer / sam
In sequential file stage, properties tab --- options -- filter
At filter type unix command -- sed '1d;$d'
| Is This Answer Correct ? | 8 Yes | 0 No |
In order to exclude the contents while reading from the
sequential file stage we have to use the filter option.
There we have to specify the unix command:
Sed -e '1d' -e '$d' <filename>
| Is This Answer Correct ? | 8 Yes | 1 No |
In sequential file stage use the filter in the options
In filter use sed '1d;$d'
You can use sed '<start>,<end>d' to delete particular rows and use sed -n '<start>,<end>p' to keep particular rows
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / shiv
Sreenu is also right, but we have another option as well.
Use head and tail stages in datastage like this:
Seq_File-->Tail_stage--->Head_Stage---->Dataset(Target)
in this way first u can exclude the first record using tail specify 999 record in tail stage, then u can exlclude last record by using head stage.(998 top records).
| Is This Answer Correct ? | 1 Yes | 2 No |
Well if must use filter option in sequential stage then
the Ans is:
head -n 10 'FileName' | tail -n 8
explanation:
suppose if ur file has 10 records + column name then total
lines are 11.
col1
1
2
3
4
5
6
7
8
9
10
head -n 10 will fetch column name + 9 records
and
tail -n 8 will fetch 2 to 9 records excluding colname and
1st record.
like :
2
3
4
5
6
7
8
9
Thats it.
| Is This Answer Correct ? | 0 Yes | 2 No |
I have load a Dataset in UAT with 2 Node configuration, imported the job into PROD environment which is 4 node configuration and using this DataSet as SRC to other job. will the job run fine or give any errors? If job runs fine, on how many nodes? 2 nodes or 4 nodes?
what should be ensure to run the sequence job so that if its get aborted in 10th job before 9job should get succeeded?
What is confirmed Dimension? what is Factless Fact? give one example? What are Additive, Semi-Additive Facts?
i have a small question for datastage, After the desinging (i.e., transformations and loading)part, what we can do?
How to RD using transformer?
A flat file contains 200 records. I want to load first 50 records at first time running the job, second 50 records at second time running and so on, how u can develop this job?
1)s.key generate 1 to 700 records today. tomorrow another 400 will updated how to update the records using s.key generator? 2)source is like :-- DB --> T/F stage1 --> seq1file T/f 1 is linking with T/F2 ---> seq 2 how to load the data? in source i given some conditions those r going in seq1. The another data will going to seq2 how to do this ?
How can we select pr retrieve the particular row in dataset by using orchadmin command?
Can we use Round Robin for aggregator?is there any benefit underlying?
parallel jobs run on cluster machines. server jobs run on smp and mpp. What do mean by cluster mechines and SMP and MPP..?
If you want to use the same piece of code in different jobs, how will you achieve it?
How do you schedule or monitoring the job?