First, let me state that for our data mart we are NOT using any type of an ETL product or toolkit and are instead using custom stored procedures. The decision to go this route was one of cost and expediency, but may be "rethunk" in the future.
My question is, how fast is fast when it comes transforming millions of records from an operational system to a Data Mart star schema?
Our home-grown SPs are transforming and loading operational records at a speed of about 2500 recs/minute (just over 40 records per second).
Is that slow, is that fast? This is running on a pretty beefed up SQL 2K box, and the raw operational records are already within the DM server in a flat stage table. The process pretty much looks ups previously inserted dimension f-keys and inserts facts.
The star schema is one of our more complex and has about 10 dimensions, several of which need to be looked up. We have tried using a cursor and a simple while loo
Should we expect to be able to get substantially faster throughput with an ETL package and if so, what manitude?
Thanks,
TR
My question is, how fast is fast when it comes transforming millions of records from an operational system to a Data Mart star schema?
Our home-grown SPs are transforming and loading operational records at a speed of about 2500 recs/minute (just over 40 records per second).
Is that slow, is that fast? This is running on a pretty beefed up SQL 2K box, and the raw operational records are already within the DM server in a flat stage table. The process pretty much looks ups previously inserted dimension f-keys and inserts facts.
The star schema is one of our more complex and has about 10 dimensions, several of which need to be looked up. We have tried using a cursor and a simple while loo
Should we expect to be able to get substantially faster throughput with an ETL package and if so, what manitude?
Thanks,
TR