Hello all,
So I’ve been around BO for many years but have on recently ventured in to the Data Service space and I have come up with an issue that I do hope someone may help me with as I going bald by the minute tearing what hair I left in my head out.
So whats the issue. And I apologise now for the long note, but this may save you coming back and asking more quesions.
We are using DS 4.1 running on a Windows VM in Dev, System Test, UAT and Production.
I have create a Delta data flow to process CDC actions of Insert, Update and Delete.
The job does the usual stuff of getting all the extra information from support tables within a Query.
From that query, I split to three more queries, one for each of my CDC actions. After the Delete query, I do a Map Operation setting both Normal and Delete to Delete. This then goes of to the Target table and deletes the records.
For the Insert and Update actions, each does a little processing and then I filter these into a Merge operation. I then have another Query to sort the data on the Key. The next step is a Table Comparison which I check the box that the input may contain duplicate keys. From here its off to the Target table. No Key Generation is used as this is a Fact table and only our DIMs have keys.
So hopefully that sounded quite straight forward and it should work. Well it does in every envirionment except Production. At first I thought that maybe this was a data volume issue but its not as one of the jobs that uses this process failed last night having processed 1800 records where the night before it processed 12000 records and completed successfully. (I shouldn’t use the word failed because it actually hasn’t failed, it just hasn’t stopped).
So go back a day and I’ve made some changes to thefirst job that has not been completing.
What I’ve done is on the Insert and Update streams is to remove the Merge and Sort steps, put the Sort in the Insert and Update Queries, given each of these queries there own Table Comparison and also their own Target table.
I ran this yeasterday using Production as my input source and a copy of the Target table was put into my Development schema and I used this as the output.
Job ran in 9 minutes, happy days, put the change into Prodcution, restarted the job that had not completed and it worked perfectly. Took longer to run, but it worked.
Move forward to today. One WF earlier in the process which has worked perfectly since going to Production has now got the same problem. (The one I mentioned earlier with only updating 1800 rows). I thought fantastic, another one, but I have a fix for this. I’ve applied the fix, run the job in Dev and it completed in 4min.
Now here is the problem. I’ve gone and decided to test the WF that I fixed yesterday and it doesn’t work. I believe that it has processed every record it just doesn’t ever get to the completed status and its doing my head in. I feel much like this guy
Has anybody had this situation before, I am really lost, there are other Data Servies people that work for the company, but as this site has been a Data Stage site for many years, and there is very little quality expertise to ask.
I’m happy to look at any suggestions that you might have.
Thanks for taking the time to read this post.
plessiusa (BOB member since 2004-03-22)