Hi All,
My company is very interested in uncovering consumer web behavior that is
buried within our Apache web logs. Currently we use a product called "Sawmill"
which is resource intensive and produces limited web analytics.
We would prefer to extract this information and populate fact and dimension
tables in our data warehouse as outlined by Ralph Kimball in his book
"The Data Warehouse Toolkit" Second Edition, Chapter 14, Electronic Commerce.
Although the book outlines the final schema(s), it doesn't mention anything about
the ETL process. During our research I came across this article where Kimball
mentions a sophisticated backend ETL process. See;
So, I guess what I'm after here is, what are other people doing? How do you extract, transform and load.
What tools do you make use of? Are they any good? Would you recommend them to others?
I look forward to your feedback.
Nuffsaid.
My company is very interested in uncovering consumer web behavior that is
buried within our Apache web logs. Currently we use a product called "Sawmill"
which is resource intensive and produces limited web analytics.
We would prefer to extract this information and populate fact and dimension
tables in our data warehouse as outlined by Ralph Kimball in his book
"The Data Warehouse Toolkit" Second Edition, Chapter 14, Electronic Commerce.
Although the book outlines the final schema(s), it doesn't mention anything about
the ETL process. During our research I came across this article where Kimball
mentions a sophisticated backend ETL process. See;
So, I guess what I'm after here is, what are other people doing? How do you extract, transform and load.
What tools do you make use of? Are they any good? Would you recommend them to others?
I look forward to your feedback.
Nuffsaid.