theniteowl
Programmer
I have an unusual problem.
One of our applications produces csv files with daily job information. Every month a script concatenates all of those daily csv files into one large one and ftps it to a network share.
I have an Access database that runs queries against the csv file as a linked table to generate monthly metrics.
The problem is the application went nuts for a week or so and created between 140 and 162 duplicate entries for every job that ran through the server that week so the numbers are WAY off.
Is there any easy way for me to eliminate the duplicate records in the existing queries or will I have to setup something to process and write out new data files without the duplicates first?
As they are linked files I cannot save changes to them and they are way too large for me modify in Excel as they greatly exceed the max number of lines Excel can handle.
Any thoughts?
My alternative is to write a VBS script that will parse the file line by line looking for the duplicate randmom number ID each entry has and each duplicate has identically and write each new non-duplicate out as a text file under a new name.
Thanks.
At my age I still learn something new every day, but I forget two others.
One of our applications produces csv files with daily job information. Every month a script concatenates all of those daily csv files into one large one and ftps it to a network share.
I have an Access database that runs queries against the csv file as a linked table to generate monthly metrics.
The problem is the application went nuts for a week or so and created between 140 and 162 duplicate entries for every job that ran through the server that week so the numbers are WAY off.
Is there any easy way for me to eliminate the duplicate records in the existing queries or will I have to setup something to process and write out new data files without the duplicates first?
As they are linked files I cannot save changes to them and they are way too large for me modify in Excel as they greatly exceed the max number of lines Excel can handle.
Any thoughts?
My alternative is to write a VBS script that will parse the file line by line looking for the duplicate randmom number ID each entry has and each duplicate has identically and write each new non-duplicate out as a text file under a new name.
Thanks.
At my age I still learn something new every day, but I forget two others.