[Zope-dev] 27 million objects.
Chris McDonough
chrism@digicool.com
Thu, 26 Apr 2001 09:09:06 -0400
Erik Enge wrote:
>
> On Thu, 5 Apr 2001, Michael R. Bernstein wrote:
>
> > I'm trying to find out of there is a point where you start getting
> > non-linear performance penalties for additional objects (storing,
> > retreiving, or indexing).
>
> I've just finished adding a somewhat small number of objects: 5000.
> For every 1000th object, the Data.fs seemed to grow to about 900MB; that's
> when things started going slow, in a non-linear fashion (this is more a
> hunch than something I payed much attention to).
>
> I paused the script (fancy Unix-command: "^Z") for every 1000th object,
> packed the database (which shrunk to 19.5MB! Hmpf.) and restarted the
> script (again, fancy Unix-command: "fg"). Then I was back to the same
> speed as I initially had.
This level of growth doesn't seem like a sane level of growth... what
Zope version are you using?
>
> Does ZODB have a problem with big Data.fs files? Not that I
> know. However, I do have a really fast SCSI-subsystem here so that
> shouldn't be a big problem either.
>
> I did some copying around with a couple of gigs, and it seems that my
> hunch is right: ZODB does not have a problem with big Data.fs files, the
> hardware does.
>
> This could be caused indirectly by ZODB if it does too many operations on
> the file, but I'm not too conserned about that. Ie. a solution could be
> to have ZODB play around with the Data.fs at a less frequent pace, or do
> it in another fashion. However, that's not really solving any problems,
> unless ZODB is a total maniac with the filesystem.
>
> I'm converting to ReiserFS this afternoon, maybe that will improve things
> a bit.
>
> Someone told me that ZEO and bulk-adding could be a thing to look at...
Isn't bulk-adding what you're doing now?