cancel
Showing results for 
Search instead for 
Did you mean: 

Managing high data volumes

Former Member
0 Kudos

I would like to get some suggestions from those of you that have worked with InfoCubes containing high number of records (millions).

- Which is the maximum number of records an InfoCube can have?

- At what amount of records is recommended to split InfoCubes? (Millions, tens of millions, hundreds of millions)

- Which is the better option for splitting InfoCubes? On time basis, Company ...

- A part from Aggregation, which are the main issues to consider for performance?

Thanks in advance,

Andreu

Accepted Solutions (0)

Answers (3)

Answers (3)

Former Member
0 Kudos

Andreu,

Here is my input...

- Which is the maximum number of records an InfoCube can have?

I do not think BW has any restrictions as to how many records an Infocube can have.

- At what amount of records is recommended to split InfoCubes? (Millions, tens of millions, hundreds of millions)

With the increasing size of fact table, performance goes down. So, obviously if the number reaches 100 million, query performance will take a hit. Generally, for better performance, it would be a good design to see that cube size doesn't go beyond 20-25 million and if does, consider splitting into two.

- Which is the better option for splitting InfoCubes? On time basis, Company ...

In most of the cases, this is done on timely basis, like Cube for 2003, Cube for 2004, Cube for 2005 etc and then a multiprovider on top of these. Because most of the report requiremnts will need to see data in 1 year at a time and maintenance and archiving will be easier, as you'll be dealing with only the current year cube. But there could be other affecting factors also, depends on individual situation...

- A part from Aggregation, which are the main issues to consider for performance?

COMPRESSION will reduce the number of records and also removes the table partitions (based on requeast ID), which improves performance.

Maintain Indexes and DB stats.

Efficient query design. Do not place a lot of chars in 'rows'. Place them in 'Free Chars' instead. Users will expand if needed. Use restrictions in the 'filter' whenever possible.

restrict the use of 'NAV' attributes, if possible.

If there is any code for 'user exit vaiables' or virtual infoobjects, try to eliminate the SELECT statements or reduce the DB access if not avoidable.

Hope it helps

Gova

Former Member
0 Kudos

Hi guys,

Thank you for your inputs!

Sorry about the points ... Don't know why but I can't assign them.

Best regards,

Andreu

GFV
Active Contributor
0 Kudos

Hi Andreu,

- I have seen some Cubes with 200 milions data on ORACLE and DB2. I guess the limit depends on the DB ... Anyway I would divide data in different ICs

- On Oracle I have no problems up to 50 mil data: over some queries become too slow

- Splitting option it's up to you ... it shold be driven by a criteria that allows parallel execution: I mean IC_A for Company A and IC_B for Company B seen from a Multicube give better results with queries cross company ...

- See the docs at the following link: https://www.sdn.sap.com/sdn/developerareas/bi.sdn?page=bi_performance.htm: according to me they should be very usefull!

Hope it helps

GFV

Former Member
0 Kudos