Hi All,
I have a table say "Upd". There are 15 jobs running every 15 minutes in steps which loads data into this table as well as other tables too. The problem is that by mistake a developer loaded data without doing a lookup. Now after a while the data rows in this table increased exponentially.
INSERT UPD SELECT TableA.<COLUMN NAMES> FROM TableA LEFT JOIN UPD ON TableA.ID=UPD.ID <REST CONDITIONS> --WHERE UPD.ID IS NULL
1 row become 2, 2 becomes 4, 4 becomes 8, and so on. Now I am stuck with 250 Million rows in this table. These all are not required.
I planned to delete these rows. But the operation is taking hell lot of time. Thus I tried of another idea where I will load data before 28th March 2014 (When this fiasco started) into another table and then dropping this table(UPD). <Rest day's data loss, I am okay with>.
Thus I have created an SSIS package to do so. But here also the package is taking much time just to start at the OLEDB Source. (I mean my DFT is "Yellow" but inside it the process have not even started, assuming that OLEDB source is firstly checking the data integrity).
Could anyone please suggest what can I do here to speed up this process, or does someone have any other idea.
If any further details are needed please let me know. This is quite urgent.
Chaos isn’t a pit. Chaos is a ladder. Many who try to climb it fail and never get to try again. The fall breaks them. And some are given a chance to climb, but they refuse. They cling to the realm, or the gods, or love. Illusions. Only the ladder is real. The climb is all there is.