[ZODB-Dev] cache not minimized at transaction boundaries?
    Chris Withers 
    chris at simplistix.co.uk
       
    Mon Jan 30 16:40:12 EST 2006
    
    
  
Hi Tim,
Tim Peters wrote:
> Do:
> 
>     import ZODB
>     print ZODB.__version__
> 
> to find out.
Good to know, thanks...
>>I have a Stepper (zopectl run on steroids) job that deals with lots of
>>big objects.
> 
> Can you quantify this?
60,000 File objects of the order of 2Mb each.
> It does not do cacheMinimize().  It tries to reduce the memory cache to the
> target number of objects specified for that cache, which is not at all the
> same as cache minimization (which latter shoots for a target size of 0).
> Whether that's "sane" or not depends on the product of:
> 
>     the cache's target number of objects
> 
> times:
> 
>     "the average" byte size of an object
Ah, that'll do it, I wondered why it was only this step that was 
hurting. My guess is that our cache size settings with lots of max-sized 
PData objects lead to the RAM blowup...
...oh well, if only the ZODB cache was RAM-usage-based ratehr than 
object count based ;-)
thanks for the info!
Chris
-- 
Simplistix - Content Management, Zope & Python Consulting
            - http://www.simplistix.co.uk
    
    
More information about the ZODB-Dev
mailing list