on 01-12-2016 3:09 PM
Hi,
I have a full refresh file with 1 million records. I should load it into a table initially and extract it in a file. From the second day I should get only the inserts and deletes. I used table comparison stage to achieve this. I tested my job with 3 lakh records and the table comparison takes approximately 10 minutes. Is there any way to improve the performance of this stage.
Note: I checked there is an option where we can set the comparison method as "sorted input" instead of row by row select. For this option, is it enough if my source file alone is sorted.
The sorted input option will always give you the best performance and should always be used with the table comparison transform, where possible. The input query needs to be sorted on the primary key of the comparison table.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.
Thank you for your inputs. I just have one clarification for this stage. I have another case where I need to do a table compare on a table that is already having records, and these records are not sorted. Would the sorted input option work in this case also. I will ensure that the records are sorted based on keys in my input file
Thank you for your inputs. I just have one clarification for this stage. I have another case where I need to do a table compare on a table that is already having records, and these records are not sorted. Would the sorted input option work in this case also. I will ensure that the records are sorted based on keys in my input file
If only inserted/deleted needed , use joins rather than table comparison transform for better performance.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.
User | Count |
---|---|
95 | |
11 | |
11 | |
10 | |
9 | |
8 | |
6 | |
5 | |
4 | |
4 |
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.