Dear Listers,
Initial investigations show that our OLAP cube will source 300 million database rows initially, and incrementally add 2 million rows per week. We're considering the use of partitions to reduce the time required to build the cube, but have no experience in design recommendations or pitfalls of use.
Background:
Our approach is to save the historical data in partitions holding the data from previous *closed* financial years. ie a partitition for each of 1999-2000 , 2000-2001. Hopefully the aggregations for these partitions, although required for most queries, need to be calculated only once (as no new data for these financial years will be added).
There will be an *active* partition to hold the changing & newly created data. This represents data yet to be closed off - it will be larger than the other partititions, holding approx 18 months of data. On a six monthly basis, we'd like to subdivide this partition, creating another *closed* partition and re-create the active partition from the significantly reduced data set.
Questions :
* are partitions reliable?
* is the proposed strategy a reasonable way to approach the problem or is there a more effective way?
* the Analysis Services manual cautions of " incorrect handling of partitions can give the cube incorrect data". Are partitions difficult to use?
* is the cube performance (query response) hindered in any way by this strategy?
* can you build or rebuild a partition independantly of the others?
* merging partitions appears to be supported, is there any difficulty in subdividing partititions?
TIA
John
Initial investigations show that our OLAP cube will source 300 million database rows initially, and incrementally add 2 million rows per week. We're considering the use of partitions to reduce the time required to build the cube, but have no experience in design recommendations or pitfalls of use.
Background:
Our approach is to save the historical data in partitions holding the data from previous *closed* financial years. ie a partitition for each of 1999-2000 , 2000-2001. Hopefully the aggregations for these partitions, although required for most queries, need to be calculated only once (as no new data for these financial years will be added).
There will be an *active* partition to hold the changing & newly created data. This represents data yet to be closed off - it will be larger than the other partititions, holding approx 18 months of data. On a six monthly basis, we'd like to subdivide this partition, creating another *closed* partition and re-create the active partition from the significantly reduced data set.
Questions :
* are partitions reliable?
* is the proposed strategy a reasonable way to approach the problem or is there a more effective way?
* the Analysis Services manual cautions of " incorrect handling of partitions can give the cube incorrect data". Are partitions difficult to use?
* is the cube performance (query response) hindered in any way by this strategy?
* can you build or rebuild a partition independantly of the others?
* merging partitions appears to be supported, is there any difficulty in subdividing partititions?
TIA
John