Avoiding using windows alltogether is the goal (sounds familiar?). The idea is to have a script on the unix machine to download the mdb file from the internet then process it. Currently the preliminary export to txt on windows is an overhead and requires a manual intervention.
Hi,
Is there any way to convert an MS Access file before doing some processing on a Unix system ? Tried using strings but some information gets lost, particularly dates.
Thanks for help
In order to compare the different solutions i've been using a big file which is not in the average. It has 39,359,150 lines (1.1G / 1132930175 bytes). Using a smaller test file may not have provided a clear cut answer about the performance.
Yep you're entirely correct, there is no conversion at all taking place. It's definitely not the best solution but it'll do until i can get my hand on some proper documentation.
Cheers
I wish i could do anything about the data types. It's just given, can't change it. This format is indeed providing dates with a precision to the tenth of second and the time is UTC. What comes after the + sign is the offset from GMT. And "substr" is not available .. a pain to have such limited...
Thanks Hoinz. Your suggestion runs indeed fine in Oracle but doesn't in the timesten table:
Yes the T is always there. Unfortunately what comes after the "." changes so apart from the T and the punctuation everything else is variable in that column.
Command> select...
Forgot to add: i can't use substr to just take the date - as i don't actually need the time part that comes after the "T" character-; that's because this has to run in a timesten database which has limited oracle functionalities.
Hi,
How do i convert a string like 2007-14-11T11:45:47.0+0100 into a date ?
I've tried the follwoing:
SQL> select to_date(f1,'yyyy-dd-mmTHH24:MI:SS.S+0100') from mx_tmp;
select to_date(f1,'yyyy-dd-mmTHH24:MI:SS.S+0100') from mx_tmp
*
ERROR at line 1:
ORA-01821: date format...
The results are nearly equivalent (yes it's solaris). Here's how a record really is if you'd like to give it a try. A would be "calling_number", B "dialled_number" and C "EL_ANM_TIMEPOINT_RECEIVED"
RECORD
#input_id 1202842387x001_23
#output_id
#input_type end_of_call_cdb
#output_type B2B_CDR...
Guess it must be the amount of data to filter out that makes the difference. The results above are obtained with "real" data, containing lots of unwanted rows. I've run tests similar to yours (test file obtained by concatenating the first record until over 1M lines are reached), here are the...
Hi,
The input to be filtered out can be considered as a stream. Well, in reality the routine produces files which have over 1M lines in average. And there's always lots of them (100 000's at a time). The benchmarking with a single file gives the following:
grep+paste -> 30 sec.
awk regexp ->...
Hi all,
With a routine producing a flow like this
header 01
0 X 52
0 G 78
0 T 44
0 B 42
0 Q 70
..
0 A 77
..
0 C 48
0 M 67
0 F 12
..
trailer 01
header 02
0 M 24
0 R 45
..
0 A 45
..
0 C 36
..
trailer 02
etc.
the desired output is like that:
77,42,48
45,,36
etc.
It's basically an extraction of...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.