Hi
I have a large file approx. 10 million records or 600mb.
I would like to remove dupicates based on the first 10
fields. I have a script which builds an array to match
against which works fine on smaller files but my pc falls
over if I try to do the same thing with large files.
Is there a better way to proccess a large file like this with awk?
thanks
I have a large file approx. 10 million records or 600mb.
I would like to remove dupicates based on the first 10
fields. I have a script which builds an array to match
against which works fine on smaller files but my pc falls
over if I try to do the same thing with large files.
Is there a better way to proccess a large file like this with awk?
thanks