Suppose we have a csv file with contents as below which is
used has a source file.
C1,C2,C3,C4
100,200,A B,300
200,400,X,Y,299 ---> it should be 4 values but by mistake
300,600,C D,566 with a comma in between X & Y,its 5.
My target takes only 4 columns, How to take care of the 2
record as above without rejecting it and loading in the target.
Hope u guys have understood my scenerio
Answers were Sorted based on User's Feedback
Answer / abhinav
I am not sure if the solution I am going to post will
actually work but it is worth a shot:
1. Don't define a delimiter so that the records are read
as '100,200,A B,300' and '200,400,X,Y,299' and so on i.e.
as a single record. This is to ensure that all the records
pass the source-qualifier.
2. Now you have all the records including the faulty one in
SQ. Next, you put an expression where you use REPLACECHAR
to remove all the commas from all the records. Now ALL your
records will of the same length.
3. Next you can use SUBSTR to read the first 3 numbers to
get the first column, next 3 to get the second column and
so on and so forth to get proper data in proper columns.
My explanation might not be clear but I have a hunch that
this might work.
| Is This Answer Correct ? | 5 Yes | 4 No |
Answer / dj
ou can do this by executing a pre session unix command. Replace consecutive delimiters with some Junk value and later in informatica mapping replace it with null.
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / naveen
By using replace char function we can write this
Suppose
length(replacechr(c3,'1234567890abcdefghtijklmnopqrstuvwzyz',null))
| Is This Answer Correct ? | 0 Yes | 1 No |
1)you have multiple source system where u receive files ,how do you actually load into mapping using transformation,what are the transformation you use? 2)you have files in ftp location ,how do you get it into mapping with you ETL concept?
What is joiner change?
i have two flat files.. containing same type of data i want to load it to dwh..how many source qualifires i need
Explain sessions?
Hi friends I want to know about what r the dimensions in the banking porjects and genarally how many tables r in project ?
I've a denormalized table EMP(empno,ename,phone,addr1,addr2,fax,email). Now I want to insert the record into target in the format that, there will be 2 target rows for 1 source record. In target the row needs to be inserted as: ROW1: empno,ename,addr1,phone ROW2: empno,ename,addr2,fax,email And I'm not allowed to normalize transformation. How to achieve?
to improve the performance of aggregator we use sorted input option and use sorter t/r befor aggregator. But here we are increasing one more cache in our mapping i.e; sorter. So how can u convince that you are increasing the performance.?
What is the difference between Active and Passive transformation?
What if i will group by with some port in aggregator and will not pass the sorted values. Will the session fails tell in both cases ( if i have configured it for sorted input and if i have not configured it for sorted input).
source file name xyz a,0,a,a,a b,b,b,0,b c,c,c,0,c target should be like this xyz a b c how to implement this?
I have table with ID,PRD_DT,PRD_FLAG,CUST_DT,CUST_FLAG I need to get max date and its corresponding flag for both the date columns. Ex:- 1A,10/3/2015,AC,10/3/2015,XY 1A,10/4/2015,AB,10/2/2015,XZ Output needed 1A,10/4/2015,10/3/2015,XY I have 100 million + in the table so avoid self-join...
I am Unable to load the FixedWith FlatFile Into The Target.What Is The Reason. PLZ Help Me..