I see such large amounts of memory here as well -- when I have a 60M database and just two users (it's our demo machine). The OODB contains 30 directories of 20 files each of 100 KB text. When we fire up /manage and peek into one of those dirs, <wham> up goes the load, until I restart the server and then it drops down to nothing. Oddly, accessing the site via FTP seems to place little load, so I presume the Zope persistence mechanism is reading into memory the entire content of each 100 KB object, just to present it in a list on the manage screen. (?) We're looking at switching to placeholder Zope objects and storing the text as static web files outside Zope. -Jeff Rush On Tue, 19 Oct 1999 02:16:49 +1000, Ben Leslie wrote:
Whilst my site is not get huge amounts of hits (yet :). I have been needing to suck the content from using wget. When I do however my system really starts to thrash badly and checking the top output I get this:
<snip> MEM CPU% MEM% 30200 www-data 6 0 60328 56M 872 R 0 13.2 45.5 1:40 python 30212 www-data 0 0 60328 56M 872 S 0 0.0 45.5 0:00 python 30213 www-data 4 0 60328 56M 872 R 0 9.0 45.5 3:11 python 30214 www-data 1 0 60328 56M 872 S 0 10.4 45.5 3:09 python 30215 www-data 8 0 60328 56M 872 R 0 18.7 45.5 3:01 python 30216 www-data 10 0 60328 56M 872 R 0 19.1 45.5 3:14 python <snip>
My site is only medium size (~200 pages, 16MB), however as you see it seems to use a fiar bit of memory, is it normal for it to use so much memory even thought my data.fs is 16mb (packed).