We are testing stress testing SER with location and subscriber databases containing 500K to 1 million entries through an artificially generated workload.

 

With the stock SER configuration when we stop and restart SER and it runs out of memory almost immediately upon startup.

It runs out of memory in convert_rows() at around row #2740 out of many (> 500000) rows.

 

If we bump up PKG_MEM_POOL_SIZE from 1024*1024 to 100*1024*1024 ser starts up successfully. However, my concern is that the memory pool is allocated in each of the 20 or 30 children, sucking up a lot of resources.

 

So, my question is this: why isn't this memory allocation truly dynamic rather than pre-allocated at startup? Or is there some better solution that I am missing completely?

 

We also have similar issues with shared memory, and need to bump this allocation via the -m command line option to ser.

 

FWIW, we are running SER 0.8.14 with a couple of local modifications.

 

Any advice would be appreciated

 

Thanks.

 

Bruce Bauman

Sr. Principal S/W Engineer

WorldGate Communications, Inc.

3190 Tremont Avenue

Trevose, PA 19053

 

Office: 215.354.5124

Cell:    215.768.8613